CN112529770A - Image processing method, image processing device, electronic equipment and readable storage medium - Google Patents
Image processing method, image processing device, electronic equipment and readable storage medium Download PDFInfo
- Publication number
- CN112529770A CN112529770A CN202011414651.7A CN202011414651A CN112529770A CN 112529770 A CN112529770 A CN 112529770A CN 202011414651 A CN202011414651 A CN 202011414651A CN 112529770 A CN112529770 A CN 112529770A
- Authority
- CN
- China
- Prior art keywords
- input
- dimensional
- dimensional model
- image
- image processing
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000012545 processing Methods 0.000 title claims abstract description 42
- 238000003672 processing method Methods 0.000 title claims abstract description 42
- 238000006073 displacement reaction Methods 0.000 claims description 35
- 238000000034 method Methods 0.000 description 17
- 230000006870 function Effects 0.000 description 13
- 230000004044 response Effects 0.000 description 9
- 238000010586 diagram Methods 0.000 description 8
- 230000008569 process Effects 0.000 description 8
- 230000008859 change Effects 0.000 description 6
- 230000004048 modification Effects 0.000 description 6
- 238000012986 modification Methods 0.000 description 6
- 238000004891 communication Methods 0.000 description 5
- 230000000694 effects Effects 0.000 description 4
- 238000007726 management method Methods 0.000 description 4
- 238000005259 measurement Methods 0.000 description 3
- 230000009286 beneficial effect Effects 0.000 description 2
- 238000001514 detection method Methods 0.000 description 2
- 230000006872 improvement Effects 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 238000005316 response function Methods 0.000 description 2
- 238000004904 shortening Methods 0.000 description 2
- 238000004364 calculation method Methods 0.000 description 1
- 230000001427 coherent effect Effects 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000007599 discharging Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000003702 image correction Methods 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000010295 mobile communication Methods 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000008439 repair process Effects 0.000 description 1
- 230000002441 reversible effect Effects 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 230000005236 sound signal Effects 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/08—Projecting images onto non-planar surfaces, e.g. geodetic screens
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/20—Editing of 3D images, e.g. changing shapes or colours, aligning objects or positioning parts
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Architecture (AREA)
- Computer Graphics (AREA)
- Computer Hardware Design (AREA)
- General Engineering & Computer Science (AREA)
- Software Systems (AREA)
- Processing Or Creating Images (AREA)
Abstract
The embodiment of the application provides an image processing method and device, electronic equipment and a readable storage medium, and belongs to the technical field of image processing. The image processing method comprises the following steps: acquiring first depth-of-field information of a target image; projecting a three-dimensional model of a target image in a space where the electronic equipment is located according to the first depth-of-field information; receiving a first input of a three-dimensional model target position; and responding to the first input, and adjusting the three-dimensional size of the target position of the three-dimensional model according to the input parameters of the first input. Therefore, the size of an object in the image can be accurately identified through the depth of field information of the image, the three-dimensional model is projected in the space, a user can modify the value of the three-dimensional size of the three-dimensional model through the operation on the target position of the three-dimensional model, the three-dimensional editing processing is realized, people or objects in the image are more vivid and stereoscopic, the delicate and perfect image is obtained, and the satisfaction degree of the user on the processed image is effectively improved.
Description
Technical Field
The present disclosure relates to the field of image processing technologies, and in particular, to an image processing method, an image processing apparatus, an electronic device, and a readable storage medium.
Background
In the related art, the method for modifying the depth image can only be operated on a plane, the modification cannot be performed according to the distance of an object and the size of a person, the problem that the modified depth image has inconsistent size or insufficient refinement exists, and the editing of the depth image on the plane cannot realize the image editing operation, so that the overall stereoscopic impression and refinement of the picture are not facilitated.
Disclosure of Invention
The embodiment of the application provides an image processing method, an image processing device, electronic equipment and a readable storage medium, which can accurately identify depth of field information in an image, project a three-dimensional model in space, change two-dimensional edition into three-dimensional edition through the three-dimensional model, and enable people or static objects in the image to be more three-dimensionally refined.
In order to solve the above problems, the present application is implemented as follows:
in a first aspect, an embodiment of the present application provides an image processing method, including:
acquiring first depth-of-field information of a target image;
projecting a three-dimensional model of a target image in a space where the electronic equipment is located according to the first depth-of-field information;
receiving a first input of a three-dimensional model target position;
and responding to the first input, and adjusting the three-dimensional size of the target position of the three-dimensional model according to the input parameters of the first input.
In a second aspect, an embodiment of the present application provides an image processing apparatus, including:
the acquisition module is used for acquiring first depth-of-field information of the target image;
the projection module is used for projecting a three-dimensional model of the target image in the space where the electronic equipment is located according to the first depth of field information;
the receiving module is used for receiving a first input of a target position of the three-dimensional model;
and the processing module is used for responding to the first input and adjusting the three-dimensional size of the target position of the three-dimensional model according to the input parameters of the first input.
In a third aspect, an embodiment of the present application provides an electronic device, which includes a processor, a memory, and a program or instructions stored on the memory and executed on the processor, where the program or instructions, when executed by the processor, implement the steps of the image processing method as provided in the first aspect.
In a fourth aspect, the present application provides a readable storage medium, on which a program or instructions are stored, which when executed by a processor implement the steps of the image processing method as provided in the first aspect.
In a fifth aspect, an embodiment of the present application provides a chip, where the chip includes a processor and a communication interface, where the communication interface is coupled to the processor, and the processor is configured to execute a program or instructions to implement the steps of the image processing method as provided in the first aspect.
In the embodiment of the application, first depth information of a target image is acquired; projecting a three-dimensional model of a target image in a space where the electronic equipment is located according to the first depth-of-field information; receiving a first input of a three-dimensional model target position; and responding to the first input, and adjusting the three-dimensional size of the target position of the three-dimensional model according to the input parameters of the first input. Therefore, the size of an object in the image can be accurately identified through the depth of field information of the image, the three-dimensional model is projected in the space, a user can modify the three-dimensional size of the three-dimensional model through the operation on the target position of the three-dimensional model, the three-dimensional editing processing is realized, people or objects in the image are more vivid and stereoscopic, the delicate and perfect image is obtained, and the satisfaction degree of the user on the processed image is effectively improved.
Drawings
FIG. 1 shows one of the flow diagrams of an image processing method according to one embodiment of the present application;
FIG. 2 shows a second flowchart of an image processing method according to an embodiment of the present application;
FIG. 3 shows a third flowchart of an image processing method according to an embodiment of the present application;
FIG. 4 shows a fourth flowchart of an image processing method according to an embodiment of the present application;
FIG. 5 shows a fifth flowchart of an image processing method according to an embodiment of the present application;
FIG. 6 shows a sixth flowchart of an image processing method according to an embodiment of the present application;
FIG. 7 shows a seventh flowchart of an image processing method according to an embodiment of the present application;
FIG. 8 shows an eighth flowchart of an image processing method according to an embodiment of the present application;
FIG. 9 shows a Gaussian curve diagram of a depth image in accordance with an embodiment of the present application;
fig. 10 shows one of the configuration block diagrams of an image processing apparatus according to an embodiment of the present application;
fig. 11 shows a second configuration block diagram of an image processing apparatus according to an embodiment of the present application;
fig. 12 shows a third block diagram of the configuration of an image processing apparatus according to an embodiment of the present application;
FIG. 13 shows a block diagram of an electronic device according to an embodiment of the present application;
fig. 14 shows a hardware configuration block diagram of an electronic device according to an embodiment of the present application.
Detailed Description
In order that the above objects, features and advantages of the present application can be more clearly understood, the present application will be described in further detail with reference to the accompanying drawings and detailed description. It should be noted that the embodiments and features of the embodiments of the present application may be combined with each other without conflict.
In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present application, however, the present application may be practiced in other ways than those described herein, and therefore the scope of the present application is not limited by the specific embodiments disclosed below.
An image processing method, an image processing apparatus, an electronic device, and a readable storage medium according to some embodiments of the present application are described below with reference to fig. 1 to 14.
In an embodiment of the present application, fig. 1 shows one of flowcharts of an image processing method of the embodiment of the present application, including:
102, acquiring first depth of field information of a target image;
for example, the mobile phone enters an album editing interface, and the user clicks an image selection target image in the album and calls first depth-of-field information recorded when the image is shot.
104, projecting a three-dimensional model of a target image in a space where the electronic equipment is located according to the first depth of field information;
in this embodiment, the first depth of field information of each pixel point of the target image is read, and the three-dimensional size of an object or a person in the target image, that is, the coordinates of the pixel points (X, Y, Z axis coordinates), is obtained according to the first depth of field information, and a three-dimensional model corresponding to the target image is projected in a space where the electronic device is located by using a plurality of projection devices in different directions according to the coordinates of the pixel points, so that a user can view the three-dimensional form of the object or the person in the target image through the three-dimensional model, and the user can select a target position to be edited and modified conveniently.
It is understood that after the three-dimensional model is projected in the space, the three-dimensional model can be synchronously displayed on the screen of the electronic device, as shown in fig. 9, and the depth information of the target image is reflected by a gaussian curve.
the image processing method is suitable for electronic equipment, and the electronic equipment comprises but is not limited to a mobile terminal, a tablet computer, a notebook computer, a wearable device, a vehicle-mounted terminal and the like. The first input may be an operation of the electronic device by the user, or an operation of the three-dimensional model recognized by the electronic device by the user. Wherein the first input includes, but is not limited to, a click input, a key input, a fingerprint input, a swipe input, and a press input. The key input includes, but is not limited to, a power key, a volume key, a single-click input of a main menu key, a double-click input, a long-press input, a combination key input, etc. to the electronic device. The operation mode in the embodiments of the present application is not particularly limited, and may be any realizable mode.
It should be noted that a photosensitive element array composed of a plurality of photosensitive elements is arranged in a space where the electronic device is located, luminance data of different positions of the three-dimensional model can be collected through the photosensitive element array, when the user operates to shield projection light beams of the three-dimensional model, the luminance data can be collected through the plurality of photosensitive elements to determine the position of the user operation, and then the target position of the three-dimensional model is identified.
For example, a finger of a user is placed at a position of the three-dimensional model to be edited, and the projection position of the finger, namely a part of pixel points in the three-dimensional model, is sensed according to the photosensitive element, so that a certain region in the image can be locally modified, three-dimensional stereo editing processing is realized, the user can edit any position in the image, the local image repairing requirement of the user is met, and the image repairing accuracy is greatly improved.
And step 108, responding to the first input, and adjusting the three-dimensional size of the target position of the three-dimensional model according to the input parameters of the first input.
In this embodiment, the three-dimensional size of the target position of the three-dimensional model is replaced or modified according to the input parameters of the first input of the target position of the three-dimensional model, namely, the correction value of the image by the user, and the modified image is stored. Therefore, the two-dimensional editing of the image is changed into the three-dimensional editing, so that the electronic equipment can carry out the three-dimensional editing operation, people or objects in the image are more vivid and stereoscopic, a more exquisite and perfect image is obtained, and the satisfaction degree of a user on the processed image is effectively improved.
It is worth mentioning that after the three-dimensional size of the target position of the three-dimensional model is adjusted according to the input parameters, the projected three-dimensional model is changed accordingly to obtain the modified three-dimensional model, so that the user can check the modification effect of the target image in time.
In an embodiment of the present application, fig. 2 shows a second flowchart of an image processing method according to an embodiment of the present application, and step 108, adjusting a three-dimensional size of a target position of a three-dimensional model according to a first input parameter, includes:
in this embodiment, the first input may be a slide input to the three-dimensional model, a motion start point and a motion end point of the slide input are identified, and a displacement between the motion start point and the motion end point is calculated. Wherein the displacement comprises a direction and a distance.
and step 208, adjusting the three-dimensional size according to the size change amount.
In this embodiment, the corresponding relationship between the preset displacement interval and the size variation is preset, that is, different displacement intervals indicate different size variations correspondingly. And comparing the displacement between the motion starting point and the motion ending point with a preset displacement interval, and taking the size variation corresponding to the preset displacement interval as a target image correction value specified by the first input under the condition that the displacement belongs to the preset displacement interval. Therefore, the three-dimensional size can be modified in real time according to the size variation, a user can dynamically adjust the three-dimensional size through the sliding operation of the three-dimensional model, the image is zoomed in three dimensions, the change of the three-dimensional model is sensed when the user edits the image, the influence on the attractiveness of the image due to excessive modification of the user is avoided, the image modifying difficulty is reduced, and the integral or local three-dimensional sense and the exquisite sense of the image are effectively improved.
For example, in the case of a portrait image, the user needs to process a nose, a flat head, and fingers placed on the nose or hair of the three-dimensional model to determine the target position. The nose can be pulled up by stretching operation, the hair with collapsed top can be pulled up, or the lateral wing of the nose can be contracted by shortening operation, etc. During the stretching/shortening operation, the target position on the image changes with the change in the slide input.
In an embodiment of the present application, fig. 3 shows a third flowchart of an image processing method according to an embodiment of the present application, and step 202, identifying a motion start point and a motion end point of a first input includes:
and step 306, determining a motion starting point and a motion ending point according to the motion track.
In this embodiment, the projection pixel point position of the first input of the target position by the user on the three-dimensional model is captured by the photosensitive element, and the plurality of projection pixel point positions are connected to generate the motion trail of the first input. The motion starting point and the motion end point of the first input can be identified according to the motion track so as to determine the displacement between the two points, thereby accurately identifying the size variation required by a user, further modifying the three-dimensional size of the three-dimensional model through the size variation, realizing three-dimensional editing processing, and being beneficial to improving the overall or local stereoscopic impression and delicate feeling of the image.
In one embodiment of the present application, fig. 4 shows a fourth flowchart of an image processing method of an embodiment of the present application, step 302, capturing a projection of a first input on a three-dimensional model, comprising:
and step 404, determining projection according to the position corresponding to the brightness data smaller than or equal to the preset threshold value.
In this embodiment, the photosensitive element array acquires brightness data of different positions of the three-dimensional model, when the brightness data is less than or equal to a preset threshold, it is indicated that the position is blocked, and may be a sliding operation position of a user, and a position corresponding to the brightness data less than or equal to the preset threshold is recorded as a projection position, so that a motion trajectory input by the user can be determined through a set of projections, so that a size variation required by the user can be accurately identified, and further, a three-dimensional size of the three-dimensional model can be modified through the size variation, and three-dimensional editing processing is realized.
In an embodiment of the present application, fig. 5 shows a fifth flowchart of an image processing method according to an embodiment of the present application, and step 108, adjusting a three-dimensional size of a target position of a three-dimensional model according to a first input parameter, includes:
in this embodiment, after the three-dimensional size of the three-dimensional model is identified, the numerical value of the three-dimensional size and the corresponding size threshold are displayed on the electronic device so that the user knows the current size parameters and the modifiable size range of the object or person in the target image. Therefore, users can reasonably repair the image according to the numerical value of the three-dimensional size and the size threshold, the image repairing quality is improved, and the image repairing difficulty is reduced.
It should be noted that the three-dimensional coordinate threshold may be a maximum value and a minimum value of the three-dimensional size of the three-dimensional model, or may be an equal-proportion adjustable range in the image that is reasonably set according to the requirement. Taking the figure retouching as an example, for the requirement of face slimming, the size threshold is the sum of the preset value and the preset value of the three-dimensional size of the pixel point. Therefore, the reasonable picture repairing range of the user can be prompted by displaying the size threshold, the influence on the attractive appearance of the picture caused by excessive modification of the user is avoided, and the picture repairing difficulty is favorably reduced.
And step 504, adjusting the three-dimensional size according to the target three-dimensional size value corresponding to the first input.
Wherein the first input is for inputting a target three-dimensional dimension value.
In this embodiment, the first input may be a key input to the electronic device, and the specific value of the target three-dimensional size value input by the user, that is, the distance information of the axes of the three-dimensional model coordinate system X, Y, Z, is obtained through the first input. Therefore, the value of the three-dimensional size of the target position of the three-dimensional model can be replaced according to the target three-dimensional size value, the three-dimensional editing function of the target image by the electronic equipment is realized, and the improvement of the whole or local three-dimensional effect and the delicate feeling of the target image is facilitated.
In an embodiment of the present application, fig. 6 shows a sixth flowchart of an image processing method according to an embodiment of the present application, including:
in this embodiment, the second input may be an operation of the electronic device by the user, or an operation of the three-dimensional model of the solid by the user recognized by the electronic device. Wherein the second input includes, but is not limited to, a click input, a key input, a fingerprint input, a swipe input, a press input. The key input includes, but is not limited to, a power key, a volume key, a single-click input of a main menu key, a double-click input, a long-press input, a combination key input, etc. to the electronic device. The operation mode in the embodiments of the present application is not particularly limited, and may be any realizable mode.
And step 604, responding to the second input, and projecting the three-dimensional model according to the rotation angle corresponding to the second input.
In this embodiment, after projecting the three-dimensional model of the target image in the space where the electronic device is located according to the first depth information, the user may control the three-dimensional model to rotate through a second input to the three-dimensional model. Therefore, a user can check the three-dimensional model in an all-around manner, the target position needing to be edited can be selected, the three-dimensional editing processing is further realized, people or objects in the image are more vivid and stereoscopic, a delicate and perfect image is obtained, and the satisfaction degree of the user on the processed image is effectively improved.
For example, the second input may be a key input of the electronic device by the user, where the key input indicates a specific numerical value of the rotation angle, and the three-dimensional model may be projected through the rotation angle to implement the rotation of the three-dimensional model. The second input may be a key input of the user to the electronic device, a control of the rotation angle is set on a screen of the electronic device, and the user adjusts the projection angle of the three-dimensional model by clicking the control of the rotation angle. In addition, the second input can also be sliding input of the user to the three-dimensional model, the motion trail of the second input is identified through the photosensitive element array, the rotation angle corresponding to the second input is matched through the motion trail, and the three-dimensional model is projected according to the rotation angle.
In an embodiment of the present application, fig. 7 shows a seventh flowchart of an image processing method according to an embodiment of the present application, where before acquiring the first depth information of the target image, step 102, the method further includes:
in this embodiment, the third input of the user to the at least one depth image may be an input of a finger of the user on the depth image, an input of a touch device such as a stylus on the depth image, or the like.
In this embodiment, at least one depth image is displayed on the screen of the electronic device, and the user may select a target image to be modified through a third input to the at least one depth image.
It is understood that a response function for triggering selection of the third input is defined for the electronic device in advance, the response function indicating that there is at least one rule triggering selection of the target image. And when a third input of the user to the electronic equipment is received, matching the third input with a rule for selecting the target image, and when the first input meets the rule, triggering the operation of determining the target image from the at least one depth image in response to the first input. For example, if the rule is defined as double-clicking the depth image, the depth image is taken as the target image when the user performs the operation of double-clicking the depth image. Of course, the rule may also be to click the depth image and confirm the control, press the depth image for a specified time, and the like, and the embodiment of the present application is not particularly limited.
Specifically, for example, taking sharing pictures from an album as an example, an album interface, that is, a thumbnail display interface of at least one depth picture, is displayed on a desktop of the electronic device, a user clicks a thumbnail to select the depth picture, and after selection, a selected identifier, that is, a "√ shaped mark", is displayed on the thumbnail of the picture, and in addition, the user lightly clicks the thumbnail to enter a large-picture browsing mode of a plurality of thumbnails, so that the thumbnail can be clearly viewed.
In an embodiment of the present application, fig. 8 shows an eighth flowchart of an image processing method according to an embodiment of the present application, where step 702, before displaying at least one depth image, further includes:
wherein, the degree of depth camera includes structure light camera and general camera.
in this embodiment, when the electronic device receives the fourth input, the depth camera is turned on to perform shooting by the depth camera. Wherein, the degree of depth camera includes structure light camera and general camera. The structured light camera may include a structured light projector and a structured light sensor. The structured light camera can project light spots, light slits, gratings, grids or stripes to an object to be detected by using a structured light projector, that is, the structured light can also be generated by using coherent light, grating light, diffraction light and the like. And then, the structured light sensor is adopted to acquire the structured light coding information of the measured object, for example, the coded pattern is modulated by the surface of the measured object.
Specifically, the structured light may be Infrared (IR) light.
By way of specific example, the projector includes a flash lamp or a continuous light source.
808, determining second depth-of-field information according to the structured light coding information;
in this embodiment, since the light with a certain structure is in different depth regions of the object, the structure of the acquired image changes from the original light structure, and then the change of the structure is converted into the second depth-of-field information by performing the distance measurement operation on the structured light encoded information.
For example, the structured light may be encoded by spatial encoding, such as de brui jn sequence encoding; the structured light may also be encoded in an acquisition time coding manner, such as binary coding, gray coding, etc.; the spatial encoding scheme may project only a single preset structured light encoded information, e.g. a single frame of structured light encoded pattern, and the temporal encoding scheme may project a plurality of different preset structured light encoded information, e.g. a plurality of frames of different structured light encoded patterns.
Specifically, for the spatial coding mode, after the collected structured light coded information is decoded, the structured light coded information and the preset structured light coded information are compared to obtain a matching relationship between the two, and the second depth-of-field information is calculated by combining a triangulation distance measurement principle. According to the time coding mode, the structured light sensor can collect a plurality of structured light coded information modulated by the surface of the moving object, the obtained plurality of structured light coded information are decoded, and the second depth of field information is obtained through calculation by combining a triangular distance measurement principle.
And step 810, performing three-dimensional reconstruction by using the second depth information to obtain at least one depth image.
In this embodiment, after the second depth information is obtained, the three-dimensional size (X, Y, Z-axis coordinates) of each pixel point is generated according to the second depth information of each pixel point, and then three-dimensional reconstruction is performed according to the three-dimensional size, so that a depth image can be obtained.
In one embodiment of the present application, as shown in fig. 10, an image processing apparatus 900 includes: the acquiring module 902, wherein the acquiring module 902 is configured to acquire first depth-of-field information of a target image; a projection module 904, wherein the projection module 904 is configured to project a three-dimensional model of the target image in a space where the electronic device is located according to the first depth-of-field information; a receiving module 906, wherein the receiving module 906 is used for receiving a first input of a target position of the three-dimensional model; a processing module 908, the processing module 908 is configured to adjust a three-dimensional size of the target location of the three-dimensional model according to input parameters of the first input in response to the first input.
In the embodiment, the size of an object in the image is accurately identified through the depth information of the image, the three-dimensional model is projected in the space, and a user can modify the numerical value of the three-dimensional size of the three-dimensional model through the operation on the target position of the three-dimensional model, so that the three-dimensional editing processing is realized, people or objects in the image are more vivid and stereoscopic, the delicate and perfect image is obtained, and the satisfaction degree of the user on the processed image is effectively improved.
Optionally, as shown in fig. 11, the image processing apparatus 900 further includes: the identification module 910, the identification module 910 is configured to identify a motion start point and a motion end point of the first input; a determining module 912, wherein the determining module 912 is configured to determine a displacement between the motion start point and the motion end point; under the condition that the displacement belongs to the preset displacement interval, determining the size variation corresponding to the displacement according to the corresponding relation between the preset displacement interval and the size variation; the processing module 908 is further configured to adjust the three-dimensional size according to the size change amount.
Optionally, the identifying module 910 is specifically configured to: capturing a projection of a first input on a three-dimensional model; generating a motion track of the first input according to the projection; and determining a motion starting point and a motion ending point according to the motion track.
Optionally, the identifying module 910 is specifically configured to: acquiring brightness data of the three-dimensional model; and determining projection according to the position corresponding to the brightness data smaller than or equal to the preset threshold value.
Optionally, as shown in fig. 12, the image processing apparatus 900 further includes: a display module 916, wherein the display module 916 is configured to display the numerical value of the three-dimensional size and the size threshold; the processing module 908 is further configured to adjust a three-dimensional size according to a target three-dimensional size value corresponding to the first input; wherein the first input is for inputting a target three-dimensional dimension value.
Optionally, the receiving module 906 is further configured to receive a second input to the three-dimensional model; the projection module 904 is further configured to project, in response to the second input, the three-dimensional model according to the rotation angle corresponding to the second input.
Optionally, the display module 916 is further configured to display at least one depth image; the receiving module 906 is further configured to receive a third input for the at least one depth image; the obtaining module 902 is further configured to determine a target image from the at least one depth image in response to a third input.
Optionally, the receiving module 906 is further configured to receive a fourth input to the electronic device; the image processing apparatus 900 further includes: a starting module (not shown in the figure), which is used for responding to the fourth input and starting the depth camera of the electronic equipment; the acquisition module (not shown in the figure) is used for acquiring the structured light coding information of the depth camera; the obtaining module 902 is further configured to determine second depth-of-field information according to the structured light encoding information; and performing three-dimensional reconstruction by using the second depth information to obtain at least one depth image. Wherein, the degree of depth camera includes structure light camera and general camera.
In this embodiment, when each module of the image processing apparatus 900 executes its respective function, the steps of the image processing method in any of the above embodiments are implemented, so that the image processing apparatus also includes all the beneficial effects of the image processing method in any of the above embodiments, which is not described herein again.
The image processing apparatus in the embodiment of the present application may be an apparatus, or may be a component, an integrated circuit, or a chip in a terminal. The device can be mobile electronic equipment or non-mobile electronic equipment. By way of example, the mobile electronic device may be a mobile phone, a tablet computer, a notebook computer, a palm top computer, a vehicle-mounted electronic device, a wearable device, an ultra-mobile personal computer (UMPC), a netbook or a Personal Digital Assistant (PDA), and the like, and the non-mobile electronic device may be a server, a Network Attached Storage (NAS), a Personal Computer (PC), a Television (TV), a teller machine or a self-service machine, and the like, and the embodiments of the present application are not particularly limited.
The management device of the application program in the embodiment of the present application may be a device having an operating system. The operating system may be an Android (Android) operating system, an ios operating system, or other possible operating systems, and embodiments of the present application are not limited specifically.
In one embodiment of the present application, as shown in fig. 13, there is provided an electronic device 1000 comprising: the processor 1004, the memory 1002, and the program or the instructions stored in the memory 1002 and executable on the processor 1004 are executed by the processor 1004 to implement the steps of the image processing method provided in any of the above embodiments, and therefore, the electronic device 1000 includes all the advantages of the image processing method provided in any of the above embodiments, which are not described herein again.
Fig. 14 is a block diagram of a hardware structure of an electronic device 1200 implementing an embodiment of the present application. The electronic device 1200 includes, but is not limited to: radio frequency unit 1202, network module 1204, audio output unit 1206, input unit 1208, sensors 1210, display unit 1212, user input unit 1214, interface unit 1216, memory 1218, processor 1220, and the like.
Those skilled in the art will appreciate that the electronic device 1200 may further comprise a power supply (e.g., a battery) for supplying power to the various components, and the power supply may be logically connected to the processor 1220 via a power management system, so as to manage charging, discharging, and power consumption management functions via the power management system. The electronic device structure shown in fig. 14 does not constitute a limitation of the electronic device, and the electronic device may include more or less components than those shown, or combine some components, or a different arrangement of components. In the embodiment of the present application, the electronic device includes, but is not limited to, a mobile terminal, a tablet computer, a notebook computer, a palm computer, a vehicle-mounted electronic device, a wearable device, a pedometer, and the like.
The processor 1220 is configured to obtain first depth information of the target image; the display unit 1212 is configured to project a three-dimensional model of the target image in a space where the electronic device is located according to the first depth information; the user input unit 1214 is used for receiving a first input of a target position of the three-dimensional model; processor 1220 is configured to, in response to the first input, adjust a three-dimensional size of a three-dimensional model target location based on input parameters of the first input.
Further, processor 1220 is also configured to capture a projection of the first input on the three-dimensional model; generating a motion track of the first input according to the projection; and determining a motion starting point and a motion ending point according to the motion track.
Further, the processor 1220 is further configured to acquire brightness data of the three-dimensional model; and determining projection according to the position corresponding to the brightness data smaller than or equal to the preset threshold value.
Further, the display unit 1212 is also configured to display the numerical value of the three-dimensional size and the size threshold; processor 1220 is further configured to adjust the three-dimensional size according to the target three-dimensional size value corresponding to the first input; wherein the first input is for inputting a target three-dimensional dimension value.
Further, the user input unit 1214 is further configured to receive a second input to the three-dimensional model; the display unit 1212 is further configured to project the three-dimensional model according to a rotation angle corresponding to the second input in response to the second input.
Further, the display unit 1212 is further configured to display at least one depth image; the user input unit 1214 is further for receiving a third input for the at least one depth image; processor 1220 is further configured to determine a target image from the at least one depth image in response to a third input.
Further, the user input unit 1214 is also used for receiving a fourth input to the electronic device; processor 1220 is further to turn on a depth camera of the electronic device in response to the fourth input; collecting structured light coding information of a depth camera; determining second depth-of-field information according to the structured light coding information; and performing three-dimensional reconstruction by using the second depth information to obtain at least one depth image.
It should be understood that, in the embodiment of the present application, the radio frequency unit 1202 may be used for transceiving information or transceiving signals during a call, and in particular, receiving downlink data of a base station or sending uplink data to the base station. Radio frequency unit 1202 includes, but is not limited to, an antenna, at least one amplifier, a transceiver, a coupler, a low noise amplifier, a duplexer, and the like.
The network module 1204 provides wireless broadband internet access to the user, such as helping the user send and receive e-mails, browse web pages, and access streaming media.
The audio output unit 1206 may convert audio data received by the radio frequency unit 1202 or the network module 1204 or stored in the memory 1218 into an audio signal and output as sound. Also, the audio output unit 1206 may provide audio output related to a specific function performed by the electronic apparatus 1200 (e.g., a call signal reception sound, a message reception sound, and the like). The audio output unit 1206 includes a speaker, a buzzer, a receiver, and the like.
The input unit 1208 is used to receive audio or video signals. The input Unit 1208 may include a Graphics Processing Unit (GPU) 5082 and a microphone 5084, and the Graphics processor 5082 processes image data of still pictures or video obtained by an image capturing device (e.g., a camera) in a video capture mode or an image capture mode. The processed image frames may be displayed on the display unit 1212 or stored in the memory 1218 (or other storage medium) or transmitted via the radio frequency unit 1202 or the network module 1204. The microphone 5084 may receive sound and may be capable of processing the sound into audio data, and the processed audio data may be converted into a format output transmittable to a mobile communication base station via the radio frequency unit 1202 in the case of a phone call mode.
The electronic device 1200 also includes at least one sensor 1210, such as a fingerprint sensor, a pressure sensor, an iris sensor, a molecular sensor, a gyroscope, a barometer, a hygrometer, a thermometer, an infrared sensor, a light sensor, a motion sensor, and others.
The display unit 1212 is used to display information input by the user or information provided to the user. The display unit 1212 may include a display panel 5122, and the display panel 5122 may be configured in the form of a liquid crystal display, an organic light emitting diode, or the like.
The user input unit 1214 may be used to receive input numeric or character information and generate key signal inputs related to user settings and function control of the electronic device. Specifically, the user input unit 1214 includes a touch panel 5142 and other input devices 5144. Touch panel 5142, also referred to as a touch screen, can collect touch operations by a user on or near it. The touch panel 5142 may include two parts of a touch detection device and a touch controller. The touch detection device detects the touch direction of a user, detects a signal brought by touch operation and transmits the signal to the touch controller; the touch controller receives touch information from the touch sensing device, converts the touch information into touch point coordinates, and sends the touch point coordinates to the processor 1220 to receive and execute commands sent by the processor 1220. Other input devices 5144 may include, but are not limited to, a physical keyboard, function keys (e.g., volume control keys, switch keys, etc.), a trackball, a mouse, and a joystick, which are not described in detail herein.
Further, the touch panel 5142 can be overlaid on the display panel 5122, and when the touch panel 5142 detects a touch operation thereon or nearby, the touch panel is transmitted to the processor 1220 to determine the type of touch event, and then the processor 1220 provides a corresponding visual output on the display panel 5122 according to the type of touch event. The touch panel 5142 and the display panel 5122 can be provided as two separate components or can be integrated into one component.
The interface unit 1216 is an interface for connecting an external device to the electronic apparatus 1200. For example, the external device may include a wired or wireless headset port, an external power supply (or battery charger) port, a wired or wireless data port, a memory card port, a port for connecting a device having an identification module, an audio input/output (I/O) port, a video I/O port, an earphone port, and the like. The interface unit 1216 may be used to receive input (e.g., data information, power, etc.) from an external device and transmit the received input to one or more elements within the electronic apparatus 1200 or may be used to transmit data between the electronic apparatus 1200 and the external device.
The processor 1220 performs various functions of the electronic device 1200 and processes data by running or executing applications and/or modules stored within the memory 1218 and by invoking data stored within the memory 1218 to thereby provide an overall monitoring of the electronic device 1200. Processor 1220 may include one or more processing units; the processor 1220 may integrate an application processor, which mainly handles operations of an operating system, a user interface, an application program, etc., and a modem processor, which mainly handles operations of image processing.
In an embodiment of the present application, a readable storage medium is provided, on which a program or instructions are stored, which when executed by a processor implement the steps of the image processing method as provided in any of the above embodiments.
In this embodiment, the readable storage medium can implement each process of the image processing method provided in the embodiment of the present application, and can achieve the same technical effect, and is not described herein again to avoid repetition.
The processor is the processor in the communication device in the above embodiment. Readable storage media, including computer-readable storage media, such as Read-Only Memory (ROM), Random Access Memory (RAM), magnetic or optical disks, etc.
The embodiment of the present application further provides a chip, where the chip includes a processor and a communication interface, the communication interface is coupled to the processor, and the processor is configured to execute a program or an instruction to implement each process of the embodiment of the image processing method, and the same technical effect can be achieved.
It should be understood that the chips mentioned in the embodiments of the present application may also be referred to as system-on-chip, system-on-chip or system-on-chip, etc.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element. Further, it should be noted that the scope of the methods and apparatus of the embodiments of the present application is not limited to performing the functions in the order illustrated or discussed, but may include performing the functions in a substantially simultaneous manner or in a reverse order based on the functions involved, e.g., the methods described may be performed in an order different than that described, and various steps may be added, omitted, or combined. In addition, features described with reference to certain examples may be combined in other examples.
Through the above description of the embodiments, those skilled in the art will clearly understand that the method of the above embodiments can be implemented by software plus a necessary general hardware platform, and certainly can also be implemented by hardware, but in many cases, the former is a better implementation manner. Based on such understanding, the technical solutions of the present application may be embodied in the form of a software product, which is stored in a storage medium (such as ROM/RAM, magnetic disk, optical disk) and includes instructions for enabling a terminal (such as a mobile phone, a computer, a server, an air conditioner, or a network device) to execute the methods of the embodiments of the present application.
While the present embodiments have been described with reference to the accompanying drawings, it is to be understood that the invention is not limited to the precise embodiments described above, which are meant to be illustrative and not restrictive, and that various changes may be made therein by those skilled in the art without departing from the spirit and scope of the invention as defined by the appended claims.
The above description is only a preferred embodiment of the present application and is not intended to limit the present application, and various modifications and changes may be made by those skilled in the art. Any modification, equivalent replacement, improvement and the like made within the spirit and principle of the present application shall be included in the protection scope of the present application.
Claims (12)
1. An image processing method, comprising:
acquiring first depth-of-field information of a target image;
projecting a three-dimensional model of the target image in a space where the electronic equipment is located according to the first depth-of-field information;
receiving a first input of a target position of the three-dimensional model;
and responding to the first input, and adjusting the three-dimensional size of the target position of the three-dimensional model according to the input parameters of the first input.
2. The image processing method according to claim 1, wherein the adjusting the three-dimensional size of the three-dimensional model target position according to the input parameters of the first input comprises:
identifying a motion start point and a motion end point of the first input;
determining a displacement between the motion start point and the motion end point;
under the condition that the displacement belongs to a preset displacement interval, determining the size variation corresponding to the displacement according to the corresponding relation between the preset displacement interval and the size variation;
and adjusting the three-dimensional size according to the size variation.
3. The image processing method of claim 2, wherein the identifying a slide start point and a slide end point of the first input comprises:
capturing a projection of the first input on the three-dimensional model;
generating a motion track of the first input according to the projection;
and determining the motion starting point and the motion ending point according to the motion track.
4. The image processing method of claim 3, wherein said capturing a projection of said first input on said three-dimensional model comprises:
acquiring brightness data of the three-dimensional model;
and determining the projection according to the position corresponding to the brightness data smaller than or equal to a preset threshold value.
5. The image processing method according to claim 1, wherein the adjusting the three-dimensional size of the three-dimensional model target position according to the input parameters of the first input comprises:
displaying the numerical value and the size threshold value of the three-dimensional size;
adjusting the three-dimensional size according to a target three-dimensional size numerical value corresponding to the first input;
wherein the first input is for inputting the target three-dimensional size value.
6. An image processing apparatus characterized by comprising:
the acquisition module is used for acquiring first depth-of-field information of the target image;
the projection module is used for projecting the three-dimensional model of the target image in the space where the electronic equipment is located according to the first depth of field information;
the receiving module is used for receiving a first input of the target position of the three-dimensional model;
and the processing module is used for responding to the first input and adjusting the three-dimensional size of the target position of the three-dimensional model according to the input parameters of the first input.
7. The image processing apparatus according to claim 6, further comprising:
the identification module is used for identifying a motion starting point and a motion ending point of the first input;
a determination module for determining a displacement between the motion start point and the motion end point; under the condition that the displacement belongs to a preset displacement interval, determining the size variation corresponding to the displacement according to the corresponding relation between the preset displacement interval and the size variation;
the processing module is further configured to adjust the three-dimensional size according to the size variation.
8. The image processing apparatus according to claim 7, wherein the identification module is specifically configured to:
capturing a projection of the first input on the three-dimensional model;
generating a motion track of the first input according to the projection;
and determining the motion starting point and the motion ending point according to the motion track.
9. The image processing apparatus according to claim 8, wherein the identification module is specifically configured to:
acquiring brightness data of the three-dimensional model;
and determining the projection according to the position corresponding to the brightness data smaller than or equal to a preset threshold value.
10. The image processing apparatus according to claim 6, further comprising:
the display module is used for displaying the numerical value of the three-dimensional size and the size threshold;
the processing module is further configured to adjust the three-dimensional size according to a target three-dimensional size numerical value corresponding to the first input;
wherein the first input is for inputting the target three-dimensional size value.
11. An electronic device comprising a processor, a memory and a program or instructions stored on the memory and executable on the processor, the program or instructions, when executed by the processor, implementing the steps of the image processing method according to any one of claims 1 to 5.
12. A readable storage medium, characterized in that it stores thereon a program or instructions which, when executed by a processor, implement the steps of the image processing method according to any one of claims 1 to 5.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011414651.7A CN112529770B (en) | 2020-12-07 | 2020-12-07 | Image processing method, device, electronic equipment and readable storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011414651.7A CN112529770B (en) | 2020-12-07 | 2020-12-07 | Image processing method, device, electronic equipment and readable storage medium |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112529770A true CN112529770A (en) | 2021-03-19 |
CN112529770B CN112529770B (en) | 2024-01-26 |
Family
ID=74997819
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202011414651.7A Active CN112529770B (en) | 2020-12-07 | 2020-12-07 | Image processing method, device, electronic equipment and readable storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112529770B (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113487727A (en) * | 2021-07-14 | 2021-10-08 | 广西民族大学 | Three-dimensional modeling system, device and method |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102055991A (en) * | 2009-10-27 | 2011-05-11 | 深圳Tcl新技术有限公司 | Conversion method and conversion device for converting two-dimensional image into three-dimensional image |
EP2347714A1 (en) * | 2010-01-26 | 2011-07-27 | Medison Co., Ltd. | Performing image process and size measurement upon a three-dimensional ultrasound image in an ultrasound system |
US20120313946A1 (en) * | 2011-06-13 | 2012-12-13 | Nobuyuki Nakamura | Display switching apparatus, display switching method, and display switching program |
CN107393017A (en) * | 2017-08-11 | 2017-11-24 | 北京铂石空间科技有限公司 | Image processing method, device, electronic equipment and storage medium |
CN108241434A (en) * | 2018-01-03 | 2018-07-03 | 广东欧珀移动通信有限公司 | Man-machine interaction method, device, medium and mobile terminal based on depth of view information |
CN108550182A (en) * | 2018-03-15 | 2018-09-18 | 维沃移动通信有限公司 | A kind of three-dimensional modeling method and terminal |
CN109727191A (en) * | 2018-12-26 | 2019-05-07 | 维沃移动通信有限公司 | A kind of image processing method and mobile terminal |
CN110908517A (en) * | 2019-11-29 | 2020-03-24 | 维沃移动通信有限公司 | Image editing method, image editing device, electronic equipment and medium |
CN111369681A (en) * | 2020-03-02 | 2020-07-03 | 腾讯科技(深圳)有限公司 | Three-dimensional model reconstruction method, device, equipment and storage medium |
-
2020
- 2020-12-07 CN CN202011414651.7A patent/CN112529770B/en active Active
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102055991A (en) * | 2009-10-27 | 2011-05-11 | 深圳Tcl新技术有限公司 | Conversion method and conversion device for converting two-dimensional image into three-dimensional image |
EP2347714A1 (en) * | 2010-01-26 | 2011-07-27 | Medison Co., Ltd. | Performing image process and size measurement upon a three-dimensional ultrasound image in an ultrasound system |
US20110184290A1 (en) * | 2010-01-26 | 2011-07-28 | Medison Co., Ltd. | Performing image process and size measurement upon a three-dimensional ultrasound image in an ultrasound system |
US20120313946A1 (en) * | 2011-06-13 | 2012-12-13 | Nobuyuki Nakamura | Display switching apparatus, display switching method, and display switching program |
CN107393017A (en) * | 2017-08-11 | 2017-11-24 | 北京铂石空间科技有限公司 | Image processing method, device, electronic equipment and storage medium |
CN108241434A (en) * | 2018-01-03 | 2018-07-03 | 广东欧珀移动通信有限公司 | Man-machine interaction method, device, medium and mobile terminal based on depth of view information |
CN108550182A (en) * | 2018-03-15 | 2018-09-18 | 维沃移动通信有限公司 | A kind of three-dimensional modeling method and terminal |
CN109727191A (en) * | 2018-12-26 | 2019-05-07 | 维沃移动通信有限公司 | A kind of image processing method and mobile terminal |
CN110908517A (en) * | 2019-11-29 | 2020-03-24 | 维沃移动通信有限公司 | Image editing method, image editing device, electronic equipment and medium |
CN111369681A (en) * | 2020-03-02 | 2020-07-03 | 腾讯科技(深圳)有限公司 | Three-dimensional model reconstruction method, device, equipment and storage medium |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113487727A (en) * | 2021-07-14 | 2021-10-08 | 广西民族大学 | Three-dimensional modeling system, device and method |
CN113487727B (en) * | 2021-07-14 | 2022-09-02 | 广西民族大学 | Three-dimensional modeling system, device and method |
Also Published As
Publication number | Publication date |
---|---|
CN112529770B (en) | 2024-01-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110675420B (en) | Image processing method and electronic equipment | |
CN111417028B (en) | Information processing method, information processing device, storage medium and electronic equipment | |
CN107592466B (en) | Photographing method and mobile terminal | |
KR102114377B1 (en) | Method for previewing images captured by electronic device and the electronic device therefor | |
CN108495029B (en) | Photographing method and mobile terminal | |
CN108712603B (en) | Image processing method and mobile terminal | |
CN108989678B (en) | Image processing method and mobile terminal | |
CN110456907A (en) | Control method, device, terminal device and the storage medium of virtual screen | |
EP2790089A1 (en) | Portable device and method for providing non-contact interface | |
CN109348135A (en) | Photographic method, device, storage medium and terminal device | |
CN108848313B (en) | Multi-person photographing method, terminal and storage medium | |
CN109510940B (en) | Image display method and terminal equipment | |
JP2015526927A (en) | Context-driven adjustment of camera parameters | |
CN111031398A (en) | Video control method and electronic equipment | |
CN108024073B (en) | Video editing method and device and intelligent mobile terminal | |
CN111541907A (en) | Article display method, apparatus, device and storage medium | |
CN109495616B (en) | Photographing method and terminal equipment | |
CN112669381B (en) | Pose determination method and device, electronic equipment and storage medium | |
CN108271012A (en) | A kind of acquisition methods of depth information, device and mobile terminal | |
CN112581571B (en) | Control method and device for virtual image model, electronic equipment and storage medium | |
CN109544445B (en) | Image processing method and device and mobile terminal | |
CN108881544A (en) | A kind of method taken pictures and mobile terminal | |
CN109246351B (en) | Composition method and terminal equipment | |
CN110908517B (en) | Image editing method, image editing device, electronic equipment and medium | |
CN111083374B (en) | Filter adding method and electronic equipment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |