CN109274976B - Data processing method and device, electronic equipment and storage medium - Google Patents

Data processing method and device, electronic equipment and storage medium Download PDF

Info

Publication number
CN109274976B
CN109274976B CN201811163098.7A CN201811163098A CN109274976B CN 109274976 B CN109274976 B CN 109274976B CN 201811163098 A CN201811163098 A CN 201811163098A CN 109274976 B CN109274976 B CN 109274976B
Authority
CN
China
Prior art keywords
pixel
data
type
coding
dimensional video
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201811163098.7A
Other languages
Chinese (zh)
Other versions
CN109274976A (en
Inventor
夏炀
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangdong Oppo Mobile Telecommunications Corp Ltd
Original Assignee
Guangdong Oppo Mobile Telecommunications Corp Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangdong Oppo Mobile Telecommunications Corp Ltd filed Critical Guangdong Oppo Mobile Telecommunications Corp Ltd
Priority to CN201811163098.7A priority Critical patent/CN109274976B/en
Publication of CN109274976A publication Critical patent/CN109274976A/en
Application granted granted Critical
Publication of CN109274976B publication Critical patent/CN109274976B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/70Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)

Abstract

The embodiment discloses a data processing method and device, electronic equipment and a storage medium. The data processing method is applied to the terminal and comprises the following steps: if the three-dimensional video data of the first type is obtained, converting the three-dimensional video data of the first type into three-dimensional video data of a second type; performing pixel coding according to the pixel value of the second type of three-dimensional video data to obtain pixel coding data; sending the pixel encoding data to a Mobile Edge Computing (MEC) server, wherein the pixel encoding data is used for restoring the three-dimensional video data by the MEC server; the data volume of the second type of three-dimensional video data before pixel coding is a first data volume; the data volume of the three-dimensional video data after pixel coding is a second data volume; the first amount of data is greater than the second amount of data.

Description

Data processing method and device, electronic equipment and storage medium
Technical Field
The present invention relates to the field of information technologies, but not limited to the field of technical information technologies, and in particular, to a data processing method and apparatus, an electronic device, and a storage medium.
Background
An image generally requires that various information such as color, gray scale, brightness, etc. of a certain pixel be represented one by a pixel value. In general, the bandwidth consumed by the image and/or video is relatively large for transmitting the same amount of information. In this way, in some image transmission scenarios, if images are continuously transmitted, a large amount of bandwidth may be consumed and/or transmission delay may be large.
Disclosure of Invention
The embodiment of the invention provides a data processing method and device, electronic equipment and a storage medium.
A data processing method is applied to a terminal and comprises the following steps:
if the three-dimensional video data of the first type is obtained, converting the three-dimensional video data of the first type into three-dimensional video data of a second type;
performing pixel coding according to the pixel value of the second type of three-dimensional video data to obtain pixel coding data;
sending the pixel encoding data to a Mobile Edge Computing (MEC) server, wherein the pixel encoding data is used for restoring the three-dimensional video data by the MEC server;
the data volume of the second type of three-dimensional video data before pixel coding is a first data volume; the data volume of the three-dimensional video data after pixel coding is a second data volume; the first amount of data is greater than the second amount of data.
Based on the above scheme, the performing pixel encoding according to the pixel values of the second type of three-dimensional video data to obtain pixel encoded data includes:
performing color coding according to the color pixel value of the second type of three-dimensional video data to obtain color coded data;
and/or the presence of a gas in the gas,
and carrying out depth value encoding according to the depth value pixel value of the second type of three-dimensional video data to obtain depth value encoded data.
Based on the above scheme, the performing pixel encoding according to the pixel values of the second type of three-dimensional video data to obtain pixel encoded data includes:
and inquiring pixel coding mapping relation according to the pixel value of the second type of three-dimensional video data, and determining the pixel coding data.
Based on the above scheme, the method further comprises:
selecting a pixel coding mode according to preset information, wherein the preset information comprises: at least one of network transmission status information, load status information of the terminal, and load status information of the MEC server;
the pixel encoding according to the pixel value of the second type of three-dimensional video data to obtain pixel encoded data includes:
and according to the selected pixel coding mode, carrying out pixel coding on the pixel value to obtain the pixel coding data.
Based on the above scheme, the pixel encoding the pixel value according to the selected pixel encoding mode to obtain the pixel encoded data includes at least one of:
according to a single-pixel coding mode, carrying out single-pixel coding on the pixel value of a single pixel of the second type of three-dimensional video data to obtain first type of coded data, wherein the number of bits occupied by the first type of coded data is smaller than the number of bits occupied by the pixel value;
and performing combined pixel coding on the pixel values of N x M pixels of the three-dimensional video data of the second type according to a combined pixel coding mode to obtain a second type of pixel coding, wherein N and M are both positive integers.
Based on the scheme, the N × M pixels are distributed adjacently;
and the N M pixels are distributed at intervals according to a preset interval mode.
Based on the above scheme, the querying a pixel coding mapping relation according to the pixel value of the second type of three-dimensional video data to determine the pixel coding data includes:
inquiring the pixel coding mapping relation according to the pixel value of the second type of three-dimensional video data;
and if the pixel value is in the pixel coding mapping relation, determining the pixel coding data according to the pixel coding value corresponding to the pixel value.
Based on the above scheme, the method further comprises:
and if the pixel value is not in the pixel coding mapping relation, updating the pixel coding mapping relation according to the pixel value, and sending the updated pixel coding mapping relation or the updated part in the pixel coding mapping relation to the MEC server.
Based on the above scheme, the performing pixel encoding according to the pixel values of the second type of three-dimensional video data to obtain pixel encoded data includes:
and obtaining the pixel value sequence number of the second type of three-dimensional video data according to the sequence of the pixel values of the second type of three-dimensional video data in a preset sequence mode.
A data processing method is applied to a Mobile Edge Computing (MEC) server and comprises the following steps:
receiving pixel coded data sent by a terminal;
restoring pixel values of the second type of three-dimensional video data according to the pixel coding data; the data volume of the second type of three-dimensional video data before pixel coding is a first data volume; the data volume of the second type of three-dimensional video data after pixel coding is a second data volume; the first amount of data is greater than the second amount of data.
Based on the above scheme, the restoring the pixel values of the second type of three-dimensional video data according to the pixel encoding data includes at least one of:
restoring color pixel values of the second type of three-dimensional video data according to color encoding data of the pixel encoding data;
and restoring the depth value pixel value of the second type of three-dimensional video data according to the depth value encoding data of the pixel encoding data.
Based on the above scheme, the method further comprises:
determining a pixel coding mode of the pixel coding data;
the restoring of the second type of three-dimensional video data from the pixel encoded data comprises:
and according to the pixel coding mode, carrying out pixel decoding on the pixel coding data to obtain a pixel value of the second type of three-dimensional video data.
Based on the above scheme, the determining the pixel encoding mode of the pixel encoding includes at least one of:
determining the number of pixels contained in the second type of three-dimensional video data, and determining the number of data of the pixel coding data; determining the pixel coding mode according to the number of the pixels and the number of the data;
interacting pixel coding parameters with the terminal, wherein the pixel coding parameters at least comprise: pixel coding method.
Based on the above scheme, the pixel decoding the pixel encoding data according to the pixel encoding mode to obtain the pixel value of the second type of three-dimensional video data includes at least one of:
decoding the pixel coding data of a single pixel to restore a second type of three-dimensional video data pixel value according to a single pixel coding mode;
and decoding the pixel coding data of the N x M pixels to restore the pixel value of the three-dimensional video data of the second type according to a combined pixel coding mode.
Based on the above scheme, the restoring the pixel values of the second type of three-dimensional video data according to the pixel encoding data includes:
and inquiring a pixel coding mapping relation according to the pixel coding data to obtain a pixel value corresponding to the pixel coding data.
Based on the above scheme, the method further comprises:
and receiving the updated pixel coding mapping relation or the updated part in the pixel coding mapping relation sent by the terminal before restoring the pixel value of the second type of three-dimensional video data according to the pixel coding data.
A data processing device is applied to a terminal and comprises:
the conversion module is used for converting the three-dimensional video data of the first type into three-dimensional video data of a second type if the three-dimensional video data of the first type is acquired;
the encoding module is used for carrying out pixel encoding according to the pixel values of the second type of three-dimensional video data to obtain pixel encoding data;
a sending module, configured to send the pixel encoding data to a moving edge computation MEC server, where the pixel encoding data is used for the MEC server to restore the second type of three-dimensional video data;
the data volume of the second type of three-dimensional video data before pixel coding is a first data volume; the data volume of the second type of three-dimensional video data after pixel coding is a second data volume; the first amount of data is greater than the second amount of data.
Based on the above scheme, the encoding module includes:
the color coding submodule is used for carrying out color coding according to the color pixel value of the second type of three-dimensional video data to obtain color coded data;
and/or the presence of a gas in the gas,
and the depth value coding submodule is used for coding the depth value according to the depth value pixel value of the second type of three-dimensional video data to obtain depth value coded data.
Based on the above scheme, the encoding module is specifically configured to query a pixel encoding mapping relationship according to the pixel value of the second type of three-dimensional video data, and determine the pixel encoding data.
Based on the above scheme, the apparatus further comprises:
the selecting module is used for selecting a pixel coding mode according to preset information, wherein the preset information comprises: at least one of network transmission status information, load status information of the terminal, and load status information of the MEC server;
the encoding module is specifically configured to perform pixel encoding on the pixel values according to a pixel encoding mode to obtain the pixel encoded data.
Based on the above scheme, the encoding module is at least configured to perform at least one of: according to a single-pixel coding mode, carrying out single-pixel coding on the pixel value of a single pixel of the second type of three-dimensional video data to obtain first type of coded data, wherein the number of bits occupied by the first type of coded data is smaller than the number of bits occupied by the pixel value; and performing combined pixel coding on the pixel values of N x M pixels of the three-dimensional video data of the second type according to a combined pixel coding mode to obtain a second type of pixel coding, wherein N and M are both positive integers.
Based on the scheme, the N × M pixels are distributed adjacently;
and the N M pixels are distributed at intervals according to a preset interval mode.
Based on the above scheme, the encoding module includes:
the query submodule is used for querying the pixel coding mapping relation according to the pixel value of the second type of three-dimensional video data;
and the determining submodule is used for determining the pixel coding data according to the pixel coding value corresponding to the pixel value if the pixel value is in the pixel coding mapping relation.
Based on the above scheme, the apparatus further comprises:
and the updating module is used for updating the pixel coding mapping relation according to the pixel value if the pixel value is not in the pixel coding mapping relation, and sending the updated pixel coding mapping relation or an updated part in the pixel coding mapping relation to the MEC server.
Based on the above scheme, the encoding module is specifically configured to obtain the pixel value sequence number of the second type of three-dimensional video data according to the ordering of the pixel values of the second type of three-dimensional video data according to a preset ordering manner.
A data processing device applied to a Mobile Edge Computing (MEC) server comprises:
the receiving module is used for receiving pixel coding data sent by a terminal;
the restoration module is used for restoring the pixel values of the second type of three-dimensional video data according to the pixel coding data; the data volume of the second type of three-dimensional video data before pixel coding is a first data volume; the data volume of the second type of three-dimensional video data after pixel coding is a second data volume; the first amount of data is greater than the second amount of data.
Based on the scheme, the reduction module comprises at least one of the following components:
the color restoration sub-module is used for restoring the color pixel value of the second type of three-dimensional video data according to the color coding data of the pixel coding data;
and the depth value restoring submodule is used for restoring the depth value pixel value of the second type of three-dimensional video data according to the depth value encoding data of the pixel encoding data.
Based on the above scheme, the apparatus further comprises:
the determining module is used for determining a pixel coding mode of the pixel coding data;
the restoring module is specifically configured to perform pixel decoding on the pixel encoded data according to the pixel encoding mode to obtain a pixel value of the second type of three-dimensional video data.
Based on the above scheme, the determining module includes at least one of:
the first determining submodule is used for determining the number of pixels contained in the second type of three-dimensional video data and determining the number of data of the pixel coding data; determining the pixel coding mode according to the number of the pixels and the number of the data;
a second determining submodule, configured to interact with the terminal for pixel encoding parameters, where the pixel encoding parameters at least include: pixel coding method.
Based on the above scheme, the reduction module is specifically configured to execute at least one of:
decoding the pixel coding data of a single pixel to restore a second type of three-dimensional video data pixel value according to a single pixel coding mode;
and decoding the pixel coding data of the N x M pixels to restore the pixel value of the three-dimensional video data of the second type according to a combined pixel coding mode.
Based on the above scheme, the restoring module is specifically configured to query a pixel coding mapping relationship according to the pixel coding data, and obtain a pixel value corresponding to the pixel coding data.
The device further comprises:
and the receiving module is used for receiving the updated pixel coding mapping relation or the updated part in the pixel coding mapping relation sent by the terminal before restoring the pixel value of the second type of three-dimensional video data according to the pixel coding data.
A computer storage medium having stored thereon computer instructions, wherein the instructions, when executed by a processor, implement the steps of any of the aforementioned data processing methods applied in a terminal; alternatively, the instructions when executed by the processor implement any of the steps of the method for data processing in the MEC server described above.
An electronic device comprising a memory, a processor and computer instructions stored on the memory and executable on the processor, wherein the processor implements any of the foregoing steps of the data processing method applied to a terminal or implements any of the foregoing steps of the data processing method applied to an MEC server when executing the instructions.
According to the method, the data processing method and device, the electronic device and the storage medium provided by the embodiment of the invention, the terminal does not directly transmit the pixel value of the second type of three-dimensional video data, but transmits the pixel encoding data after pixel encoding is carried out on the pixel value. The data volume of the transmitted pixel coding data is smaller than that of the directly transmitted pixel value, so that the bandwidth and the time delay required by transmission are reduced; the method has the characteristics of small data transmission quantity, small required bandwidth and small transmission delay.
Drawings
Fig. 1 is a schematic diagram of a system architecture of a data transmission method according to an embodiment of the present invention;
fig. 2 is a schematic flow chart of a data processing method according to an embodiment of the present invention;
FIG. 3 is a flow chart illustrating another data processing method according to an embodiment of the present invention;
fig. 4 is a schematic flowchart of another data processing method according to an embodiment of the present invention;
fig. 5 is a schematic structural diagram of a data processing apparatus according to an embodiment of the present invention;
FIG. 6 is a block diagram of another data processing apparatus according to an embodiment of the present invention;
fig. 7 is a schematic structural diagram of an electronic device according to an embodiment of the present invention.
Detailed Description
Before the technical solution of the embodiment of the present invention is explained in detail, a system architecture to which the data processing method of the embodiment of the present invention is applied is first briefly explained. The data processing method of the embodiment of the invention is applied to the related service of the second type of three-dimensional video data, such as the service of sharing the second type of three-dimensional video data, or the live broadcast service based on the second type of three-dimensional video data, and the like. In this case, since the data amount of the second type of three-dimensional video data is large, the depth data and the two-dimensional video data transmitted respectively require high technical support in the data transmission process, and thus the mobile communication network is required to have a high data transmission rate and a stable data transmission environment.
Fig. 1 is a schematic diagram of a system architecture to which a data transmission method according to an embodiment of the present invention is applied; as shown in fig. 1, the system may include a terminal, a base station, an MEC server, a service processing MEC server, a core network, the Internet (Internet), and the like; and a high-speed channel is established between the MEC server and the service processing MEC server through a core network to realize data synchronization.
Taking an application scenario of interaction between two terminals shown in fig. 1 as an example, an MEC server a is an MEC server deployed near a terminal a (a sending end), and a core network a is a core network in an area where the terminal a is located; correspondingly, the MEC server B is an MEC server deployed near the terminal B (receiving end), and the core network B is a core network of an area where the terminal B is located; the MEC server A and the MEC server B can establish a high-speed channel with the service processing MEC server through the core network A and the core network B respectively to realize data synchronization.
After the second type of three-dimensional video data sent by the terminal A is transmitted to the MEC server A, the MEC server A synchronizes the data to the business processing MEC server through the core network A; and then, the MEC server B acquires the second type of three-dimensional video data sent by the terminal A from the business processing MEC server, and sends the second type of three-dimensional video data to the terminal B for presentation.
Here, if the terminal B and the terminal a realize transmission through the same MEC server, the terminal B and the terminal a directly realize transmission of the second type of three-dimensional video data through one MEC server at this time, without participation of a service processing MEC server, and this mode is called a local backhaul mode. Specifically, it is assumed that the terminal B and the terminal a realize transmission of the second type of three-dimensional video data through the MEC server a, and after the second type of three-dimensional video data sent by the terminal a is transmitted to the MEC server a, the MEC server a sends the second type of three-dimensional video data to the terminal B for presentation.
Here, the terminal may select an evolved node b (eNB) accessing the 4G network or a next generation evolved node b (gNB) accessing the 5G network based on a network situation, or a configuration situation of the terminal itself, or an algorithm of the self-configuration, so that the eNB is connected with the MEC server through a Long Term Evolution (LTE) access network, and the gNB is connected with the MEC server through a next generation access network (NG-RAN).
Here, the MEC server is deployed on the network edge side near the terminal or the data source, that is, near the terminal or near the data source, not only in a logical location but also in a geographical location. Unlike the existing mobile communication network in which the main service processing MEC servers are deployed in several large cities, the MEC servers can be deployed in a plurality in one city. For example, in an office building, there are many users, and a MEC server may be deployed near the office building.
The MEC server serves as an edge computing gateway with the core capabilities of network convergence, computing, storage and application, and provides platform support comprising an equipment domain, a network domain, a data domain and an application domain for edge computing. The intelligent connection and data processing system is connected with various intelligent devices and sensors, provides intelligent connection and data processing services nearby, enables different types of applications and data to be processed in the MEC server, achieves key intelligent services such as real-time service, intelligent service, data aggregation and interoperation, safety and privacy protection and the like, and effectively improves intelligent decision efficiency of the service.
As shown in fig. 2, the present embodiment provides a data processing method, which is applied in a terminal and includes:
step 201: if the three-dimensional video data of the first type is obtained, converting the three-dimensional video data of the first type into three-dimensional video data of a second type;
step 203: performing pixel coding according to the pixel value of the second type of three-dimensional video data to obtain pixel coding data;
step 205: sending the pixel encoding data to a Mobile Edge Computing (MEC) server, wherein the pixel encoding data is used for restoring the second type of three-dimensional video data by the MEC server;
the data volume of the second type of three-dimensional video data before pixel coding is a first data volume; the data volume of the second type of three-dimensional video data after pixel coding is a second data volume; the first amount of data is greater than the second amount of data.
The data processing method provided by the embodiment is applied to a terminal, and the terminal can be various types of terminals, such as a mobile phone, a flat computer or a wearable device, or fixed image monitoring. The terminal may be a fixed terminal and/or a mobile terminal.
In some embodiments, the first type of three-dimensional video data and the second type of three-dimensional video data are different types of three-dimensional video data, and specifically, the first type of three-dimensional video data and the second type of three-dimensional video data may be: three-dimensional video data formed by adopting different coding modes; for example, the pixel values of the color pixels of the first type of three-dimensional video data may be: brightness/chrominance/concentration (YUV) values; the pixel values of the color pixels of the second type of three-dimensional video data may be red/green/blue (RGB) values. For another example, the pixel values of the color pixels of the first type of three-dimensional video data may be: RGB value; the pixel values of the color pixels of the second type of three-dimensional video data may be YUV values.
For example, taking an 8-bit color channel as an example, the conversion relationship between the RGB value and the YUV value may be as follows:
Y=0.299R+0.587G+0.114B
U=-0.147R-0.289G+0.436B
V=0.615R-0.515G-0.100B
R=Y+1.14V
G=Y-0.39U-0.58V
B=Y+2.03U
for example, if the pixel values of the color pixels of the first type of three-dimensional video data can be: a brightness/chrominance/density/transparency (YUVA) value, the pixel value of the color pixel of the second type of three-dimensional video data may be a red/green/blue/transparency (RGBA) value.
A represents a transparency value in the two-dimensional image. In converting YUVA to RGBA, the YUV and RGB components may be parameterized by the above functional relationship, while the a component may remain unchanged.
The depth image includes depth pixels whose pixel values are depth values. The second type of three-dimensional video data and the depth image can build a three-dimensional image in a three-dimensional image space.
The pixel values of the depth pixels may be: the absolute value of the spatial distance between the image capture device and the capture object, and/or the difference in distance between different locations on the surface of the capture object and the capture object (this may be referred to as a relative value). For example, the acquisition target is a human face, and the absolute distance between each position of the human face and the camera is the absolute value; and if the distance difference of each position of the face relative to the distance between the cameras is the relative value.
If the first type of three-dimensional video data is the absolute value of the depth pixel, the second type of three-dimensional video data is the relative value of the depth pixel; and if the second type of three-dimensional video data is the absolute value of the depth pixel, the first type of three-dimensional video data is the relative value of the depth pixel.
When converting the absolute value into a relative value, the minimum absolute value may be subtracted from the absolute value of each depth pixel to obtain the relative value. At this time, the minimum absolute value can be transmitted in addition to the pixel coding data, so that the subsequent MEC server can conveniently determine the distance of the whole three-dimensional video modeling according to the minimum absolute value. In other embodiments, the minimum absolute value may not be transmitted, and the MEC server uses a default value to determine the distance of the three-dimensional video modeling.
When the relative value is converted into the absolute value, the distance between the acquisition target and the image acquisition module can be estimated according to the two-dimensional image in the three-dimensional video data, so that the converted absolute value is obtained. Alternatively, the converted absolute value is obtained by directly adding a specific distance value to the relative value.
In summary, in the present embodiment, the first type of three-dimensional video data and the second type of three-dimensional video data include, but are not limited to, the above examples; the conversion relationship between the two types of three-dimensional video data also includes, but is not limited to, the above examples.
In this way, pixel encoding of three-dimensional data to which no pixel encoding parameter is set can be performed using pixel encoding parameters that have been configured for pixel encoding. The pixel encoding parameters include, but are not limited to, a pixel encoding mode, a pixel encoding mapping relationship, and the number and/or dimensions of pixel combinations of a combined encoding mode.
In some embodiments, the two-dimensional image and the depth image have the same image size, e.g., the two-dimensional image and the depth image each include W × H pixels; w represents the number of pixels included in the first direction, and H represents the number of pixels included in the second direction. W and H are both positive integers.
In some embodiments, the two-dimensional image and the depth image may be two images acquired at the same time; in order to reduce the data amount, the image sizes of the two-dimensional image and the depth image satisfy a preset relationship. For example, the depth image includes W × H pixels, and the depth image includes (W/a) × (H/b) pixels. Thus, one depth pixel corresponds to a × b color pixels. When a three-dimensional video is built, the pixel values of a-b adjacent color pixels can be applied according to the pixel value of one depth pixel. For example, (W/a) × (H/b) equals (W/2) × (H/2). Thus, one depth pixel corresponds to 4 color pixels. When the three-dimensional video is built, the pixel value of one depth pixel can be applied to the pixel values of 4 adjacent color pixels; in this way, the amount of image data of the depth image is reduced. Because the concave-convex feeling in the small adjacent areas of an object is basically consistent, if the image size of the depth image is smaller than that of the two-dimensional image, the high-precision three-dimensional video can be restored and built; while reducing the amount of data that the terminal and the MEC server need to interact and/or the amount of data that the MEC server needs to process. In some embodiments, when generating an image size smaller than the two-dimensional image, there may be at least one of the following:
acquiring the depth image directly by using the image size of the depth image;
acquiring an original depth image by using the image size of the two-dimensional image; and generating the depth image according to the image size of the depth image and the pixel values of the adjacent a-b pixels. For example, the depth image is generated from the mean or median of adjacent a × b pixel values.
In this embodiment, the first encoding that has been completed to convert from the sensing data to the pixel value is performed. In this embodiment, the pixel value is encoded for the second time, where the second encoding is the pixel encoding, and the pixel encoding data is obtained after the pixel encoding is completed.
In some embodiments, after pixel encoding the pixel values, the generated pixel encoded data may include: pixel value codes, not the pixel values themselves. Therefore, after receiving the pixel encoded data, the receiving end cannot directly display or read the image according to the pixel encoded data, and the receiving end needs to restore the pixel encoded data to the pixel value itself before the image can be normally displayed or read.
In some embodiments, the first encoding may be an encoding carried by many image capturing modules, so that the image capturing modules directly store pixel values of the sensing data conversion, i.e. data of the first encoding, in the storage through the capturing of light.
And transmitting the pixel coding data obtained after the pixel coding is finished to an MEC server for generating the three-dimensional video of the MEC server. In this embodiment, since the second data amount obtained after the second pixel encoding is smaller than the first data amount before the encoding, the data amount for transmitting the second type of three-dimensional video data is reduced, thereby reducing the bandwidth consumed by the data amount and the transmission delay required by a large amount of data; therefore, the method has the characteristics of small transmitted data volume, small consumed bandwidth and small transmission time delay. Therefore, the MEC server receives the data with small delay, and can quickly and accurately restore the three-dimensional video data of the second type and build the three-dimensional video.
In the embodiment of the invention, if the three-dimensional video data of the first type is obtained, the three-dimensional video data of the second type with known pixel coding parameters is converted, so that the pixel coding can be performed by using the known pixel coding parameters, and the pixel coding parameters of the three-dimensional video data of the first type with the pixel coding parameters not established yet do not need to be established.
In some embodiments, the pixel encoding according to the pixel values of the second type of three-dimensional video data to obtain pixel encoded data includes:
performing color coding according to the color pixel value of the second type of three-dimensional video data to obtain color coded data;
and/or the presence of a gas in the gas,
and carrying out depth value encoding according to the depth value pixel value of the second type of three-dimensional video data to obtain depth value encoded data.
In some embodiments, encoding the pixel values may only color encode color pixel values of color pixels in the second type of three-dimensional video data, resulting in color encoded data.
In further embodiments, the encoding of the pixel values may be depth value encoding of only depth value pixel values in the second type of three-dimensional video data, obtaining re-encoded depth value encoded data.
Regardless of color coding or depth value coding, the amount of data transmitted to the MEC server can be reduced by coding again.
In other embodiments, the pixel encoding in step 203 may be simultaneous color encoding and depth value encoding.
Step 203 may comprise:
matching pixel values in the second type of three-dimensional video data with pixel values in a pixel coding mapping relation;
and determining the pixel coding data according to the matching result. For example, the pixel value a1 in one or more of the second type of three-dimensional video data is matched with all the pixel values in the pixel coding mapping relationship, and if there is a match to the pixel value a1, the pixel coded data in the pixel coding mapping relationship corresponding to the pixel value a1 is taken as the result of the pixel coding of the pixel value a 1.
The matching result comprises the following three types:
the matching result shows that the matching is successful; the matching success includes: the matching result shows that the same condition or similar condition is satisfied;
the matching result indicates that the matching is unsuccessful; i.e. the matching results do not fulfill the same condition and/or similar conditions.
If the same conditions are met, the pixel values in the currently acquired three-dimensional video data of the second type are indicated to be in the pixel coding mapping relation.
If the similar condition is satisfied, the pixel value and the position in the currently acquired three-dimensional video data of the second type are indicated to be coded.
In some embodiments, it may be determined whether the matching success requires the same condition or a similar condition to be satisfied according to the current requirements.
In some embodiments, if the pixel degree of the pixel values of N × M pixels scanned from the currently acquired three-dimensional video data of the second type and the pixel values of a predetermined N × M pixels in the pixel coding mapping relationship is greater than a predetermined similarity threshold, for example, 70%, 80%, 90%, or 85%; then, the pixel coding data of the currently scanned N × M pixels and the N × M pixels in the pixel coding mapping relation may be considered to satisfy the similar condition of the pixel coding, and the pixel coding data of the N × M pixels in the pixel coding mapping relation may be directly used as the color coding data of the pixel values of the currently scanned N × M pixels.
In other embodiments, if the pixel degrees of the pixel values of N × M pixels in the currently acquired three-dimensional video data of the second type and the pixel values of a certain preset N × M pixels in the pixel coding mapping relationship are greater than a preset similarity threshold value; 70%, 80%, 90% or 85%. Further, extracting pixel values of 1 or more pixels with different pixel values of the N × M scanned pixels and the N × M pixels in the pixel coding mapping relation, calculating pixel value differences between the extracted pixel values and the pixel values of the N × M pixels in the pixel coding mapping relation, if the pixel value differences are within a preset difference range, considering that the pixel coding data of the N × M currently scanned pixels and the N × M pixels in the pixel coding mapping relation meet the similarity condition of pixel coding, and directly using the pixel coding data of the N × M pixels in the pixel coding mapping relation as color coding data of the pixel values of the N × M currently scanned pixels; otherwise, the scanned pixel coding data of the N × M pixels and the N × M pixels in the pixel coding mapping relation are considered not to satisfy the similar condition of the pixel coding.
For example, if the pixel value difference is within the preset difference range, the following steps may be included:
the difference in pixel values indicates that the two pixel values are approximations, e.g., color approximations. If the difference of the pixel values indicates that the two colors are opposite colors, the two colors can be considered not to be in the preset difference range any more; if the depth difference of the two depth pixels indicates that the difference between the two depth values is above the preset depth value or the depth ratio, the two depth pixels can be considered to be not in the preset difference range any more, otherwise, the two depth pixels can be considered to be in the preset difference range.
In other embodiments, if the encoding mapping relationship is an encoding mapping function, the pixel encoding data is automatically output when the pixel value is input into the encoding mapping function. For example, the encoding mapping function is determined by fitting color values in a sample image, such that each pixel value or a group of pixel values input to the encoding mapping function automatically results in the pixel encoding data, such that the pixel encoding data is not determined by means of matching.
In summary, there are various ways to determine the pixel-encoded data in step 203, and the implementation is not limited to any one.
In some embodiments, the step 203 comprises:
and inquiring pixel coding mapping relation according to the pixel value of the second type of three-dimensional video data, and determining the pixel coding data.
In some embodiments, the pixel coding mapping relationship may be known in advance by both the terminal and the MEC server, for example, a pixel coding mapping table is stored in advance by both the MEC server and the terminal.
In some embodiments, the pixel coding mapping relationship is pre-negotiated between the terminal and the MEC server.
The pixel coding mapping relationship may include at least one of:
the pixel coding mapping table;
a plurality of discrete pixel encoding map value pairs;
the data is encoded as a function of the pixel values and the pixels.
In short, the pixel coding mapping relationship has at least a plurality of expression modes, and is not limited to any one of the above.
In some embodiments, as shown in fig. 3, the method further comprises:
step 200: selecting a pixel coding mode according to preset information, wherein the preset information comprises: at least one of network transmission status information, load status information of the terminal, and load status information of the MEC server;
the step 203 may comprise: and carrying out pixel coding on the pixel values of the second type of three-dimensional video data according to the selected pixel coding mode to obtain the pixel coded data.
For example, if the network transmission status information indicates that: the pixel encoding may not be used if the currently available bandwidth is greater than the bandwidth required to directly transmit the pixel value.
For another example, if the network transmission status information indicates that: and if the current available high bandwidth is smaller than the bandwidth required for directly transmitting the pixel value, selecting a pixel coding mode of which the data volume after pixel coding is smaller than or equal to the current available bandwidth according to the current available bandwidth.
For another example, if different pixel coding modes are adopted, the calculation amount required by the terminal coding is different from the calculation amount restored by the MEC server.
In this embodiment, an appropriate pixel encoding method is also selected according to the load status information of the terminal and/or the load status information of the MEC server.
The load condition information may include at least one of: the load balancing method comprises the steps of current load rate, current load amount, maximum load rate and maximum load amount.
If the current load rate is high or the current load capacity is large, a pixel coding mode with small coding or decoding calculated amount is preferentially selected; otherwise, the selection can be arbitrarily carried out or carried out according to other reference factors such as network transmission condition information and the like.
In some embodiments, the pixel encoding the pixel value according to the selected pixel encoding mode to obtain the pixel encoding data includes at least one of:
according to a single-pixel coding mode, carrying out single-pixel coding on the pixel value of a single pixel of the second type of three-dimensional video data to obtain first type of coded data, wherein the number of bits occupied by the first type of coded data is smaller than the number of bits occupied by the pixel value;
and performing combined pixel coding on the pixel values of N x M pixels of the three-dimensional video data of the second type according to a combined pixel coding mode to obtain a second type of pixel coding, wherein N and M are both positive integers.
In the present embodiment, a single pixel is encoded, and one pixel value corresponds to one pixel encoded data. For example, if an image of the second type of three-dimensional video data includes S pixels, S pieces of the first type encoded data are obtained after encoding by a single pixel. In order to reduce the amount of data, the number of bits occupied by one of the first type encoded data is now smaller than the number of bits occupied by the pixel value itself. For example, a pixel value occupies 32 bits or 16 bits, whereas the first type of encoded data occupies only 8 bits or 10 bits. Thus, the amount of data required is reduced overall, as the number of bits required for each single pixel transfer is reduced.
Pixel coding may also be combined in some embodiments.
The combined pixel coding is pixel coding of a plurality of pixels at the same time.
For example, a matrix of adjacent pixels is encoded at the same time, or a plurality of pixels arranged in a matrix or non-matrix are encoded at the same time.
In some embodiments, a pixel matrix of 3 × 3 or 4 × 4 pixels is encoded. In some embodiments, N × M is exactly divisible by pixels included in one frame of the three-dimensional image data.
In some cases, when image capture is performed, the depth values and/or color information of these adjacent pixels are relatively fixed, these colors or depth combinations may be combined to generate a preset encoding value in the pixel encoding mapping relationship, and thus, when the pixel encoding is performed, it is determined whether a specific color combination and/or depth combination is included by scanning the color pixel values or depth pixel values in the corresponding second type of three-dimensional video data frame, so as to convert the color pixel values or depth pixel values into corresponding encoding values, thereby obtaining the pixel encoding data.
In some embodiments, the single pixel coding and the combined pixel coding may be used in a mixture, depending on the current needs.
The selected encoding mode may be informed in advance while or before the pixel encoded data is transmitted. The selected encoding method may be the single pixel encoding, the combined pixel encoding, or the mixed pixel encoding combining the single pixel encoding and the combined pixel encoding.
The N x M pixels are distributed adjacently;
and the N M pixels are distributed at intervals according to a preset interval mode.
If N M pixels are distributed adjacently, an N M pixel matrix is formed.
The N × M pixels are spaced apart by a predetermined interval, for example, two pixels belonging to the N × M pixels may be spaced apart by a predetermined number of pixels, for example, by one or more.
In some embodiments, N × M may be determined dynamically or may be set statically.
For example, an image in a three-dimensional image data frame is divided into a first region and a second region, the first region can be encoded by using a single pixel, and the second region is encoded by combining pixels.
For another example, pixel values of a first region of an image in the three-dimensional image frame are directly transmitted to the MEC server, and single-pixel coding and/or combined-pixel coding is performed on a second region.
Thus, the relationship between the transmission data amount and the image quality can be well balanced.
In some embodiments, said querying a pixel coding mapping relationship according to pixel values of said second type of three-dimensional video data, determining said pixel coding data, comprises:
inquiring the pixel coding mapping relation according to the pixel value of the second type of three-dimensional video data;
and if the pixel value is in the pixel coding mapping relation, determining the pixel coding data according to the pixel coding value corresponding to the pixel value.
The pixel coding map of the image data of a second type of frame of three-dimensional video data may have been previously determined, but in other cases may have not been determined, or may have changed over time.
For example, taking a three-dimensional live video of a main broadcast as an example, if the main broadcast participates in the three-dimensional live video before, the encoding mapping relationship of the face of the main broadcast may be stored in a terminal or an MEC server held by the main broadcast. If the anchor face is suddenly decorated or the makeup changes, at least the color image of the face may change, and the pixel mapping may not be in the pixel coding mapping.
In other embodiments, the method further comprises:
and if the pixel value is not in the pixel coding mapping relation, updating the pixel coding mapping relation according to the pixel value, and sending the updated pixel coding mapping relation or the updated part in the pixel coding mapping relation to the MEC server.
In this embodiment, in order to determine the encoding mapping relationship conveniently, one or more second types of three-dimensional video data of the target object may be acquired in an interactive handshake or debugging stage before formal live broadcasting, and whether a pixel mapping relationship corresponding to the target object has been established or not is determined or whether the pixel mapping relationship needs to be updated through pixel value scanning of the second types of three-dimensional video data. And if the three-dimensional mapping relation needs to be updated, updating the three-dimensional mapping relation, and if the three-dimensional mapping relation does not need to be updated, the formal interaction of the second type of three-dimensional video data can be directly carried out.
In some embodiments, the step 203 may comprise:
and obtaining the pixel value sequence number of the second type of three-dimensional video data according to the sequence of the pixel values of the second type of three-dimensional video data in a preset sequence mode.
For example, taking a human face as an example, the skin color and the facial height of the human face both have the maximum value and the minimum value, so that the two-dimensional image and/or the depth image acquired by the image acquisition module are concentrated in a specific color pixel value or depth pixel value interval, in most cases, the maximum pixel value and the minimum pixel value of the whole image acquisition unit are not covered, 512 possible pixel values corresponding to a 16-bit color channel are possibly effectively utilized, and only about 200, even more than 100 are possibly effectively utilized.
By the ordering of the pixel values it is possible to obtain how many pixel values are currently generated, e.g. P are generated, log is needed2The pixel coding of all pixels can be completed by taking the whole bit upwards of P, and only log is occupied2P up takes the entire pixel encoded data; thus, the required data volume can be greatly reduced.
If a target object (for example, various types of anchor, and a specific type of scene often appears in the video), the pixel coding mapping relation can be generated through the above statistical number ordering of the pixel values, or the pixel coding mapping relation is updated, so that the determination and generation of the coding video relation are completed.
In some embodiments, the obtaining the sequence number of the pixel values of the second type of three-dimensional video data according to the sorting of the pixel values of the second type of three-dimensional video data according to a preset sorting manner may include:
if a single coding mapping mode is adopted, obtaining the pixel value sequence number of the second type of three-dimensional video data according to the ordering of the pixel values of the second type of three-dimensional video data in a preset ordering mode;
and establishing a mapping relation between the pixel values and the pixel value sequence numbers.
For example, the pixel value index includes at least one of:
color value sequence numbers formed by sorting the color values;
and sorting the depth values to form depth value serial numbers.
In some further embodiments, the obtaining the pixel value sequence number of the second type of three-dimensional video data according to the sorting of the pixel values of the second type of three-dimensional video data according to a preset sorting manner may further include:
if the current coding mapping mode is a combined coding mapping mode, determining the value of N x M of the combined coding mapping mode according to the required precision and/or reference factors such as a target scene; wherein, the values of N and M are positive integers;
according to pixel values of the three-dimensional video data of the second type, the pixel values of N-M pixels are used as a combination to be sequenced, and a pixel combination serial number of the three-dimensional video data of the second type is obtained;
and establishing a mapping relation between the pixel value and the pixel combination serial number.
In this embodiment, N × M may be determined according to the required accuracy, where N may be the number of rows corresponding to one pixel combination, and M may be the number of columns corresponding to one pixel combination; alternatively, N may be the number of columns corresponding to one pixel combination, and N may be the number of rows corresponding to one pixel combination.
In some embodiments, the pixel combination order number comprises at least one of:
color value combination serial numbers formed by color value combination sorting;
and combining and sorting the depth values to form depth value combination serial numbers.
In some embodiments, the obtaining the pixel combination serial number of the second type of three-dimensional video data by sorting the pixel values of N × M pixels according to the pixel values of the second type of three-dimensional video data may include:
and sorting by taking the color values of the N × M pixels as a combination according to the color values of the color pixels in the three-dimensional video data of the second type to obtain a color value combination serial number of the three-dimensional video data of the second type. For example, the color value combinations may be sorted according to the chronological order of the scanned color value combinations, or may be sorted based on the frequency of occurrence of the scanned color value combinations, so as to obtain the color value combination serial numbers.
In some embodiments, the pixel coding mapping relation obtained by different target objects may be different, so that the pixel coding mapping relation is high in safety without being leaked for data. Therefore, if the pixel coding data subjected to pixel coding is intercepted in the transmission process, other people cannot normally decode the second type of three-dimensional video data, and therefore the transmission safety is high.
If the three-dimensional video data are sequenced according to the statistical number, the serial number of the pixel coding value with high occurrence frequency of the pixel value appears in front, so that the pixel value matching times can be reduced and the pixel coding efficiency can be improved when the three-dimensional video data of the second type of the sample and the three-dimensional video data of the second type of the collected target are coded in the same target scene.
As shown in fig. 4, the present embodiment provides a data processing method applied to a mobile edge computing MEC server, including:
step 301: receiving pixel coded data sent by a terminal;
step 303: restoring pixel values of the second type of three-dimensional video data according to the pixel coding data; the data volume of the second type of three-dimensional video data before pixel coding is a first data volume; the data volume of the second type of three-dimensional video data after pixel coding is a second data volume; the first amount of data is greater than the second amount of data.
In the present embodiment, not the pixel values but the pixel encoded data are directly received. The MEC server needs to restore the pixel values of the second type of three-dimensional video data after receiving the pixel encoding data.
Since the pixel encoded data received by the MEC server is smaller in data amount relative to the pixel value directly received, the consumed bandwidth is smaller.
In some embodiments, the step 303 may include at least one of:
restoring color pixel values of the second type of three-dimensional video data according to color encoding data of the pixel encoding data;
and restoring the depth value pixel value of the second type of three-dimensional video data according to the depth value encoding data of the pixel encoding data.
In this embodiment, the color pixel values are restored based on the color encoded data, and the depth value pixel values are restored from the depth value encoded data.
In some embodiments, the method further comprises:
step 302: determining a pixel coding mode of the pixel coding data; for example, the pixel encoding method may include: a single coding scheme and/or a combined coding scheme.
The step 303 may include:
and according to the pixel coding mode, carrying out pixel decoding on the pixel coding data to obtain a pixel value of the second type of three-dimensional video data.
In some embodiments, there are multiple ways of the step 302, and the following provides several alternatives:
the first alternative is as follows: determining the number of pixels contained in the second type of three-dimensional video data, and determining the number of data of the pixel coding data; determining the pixel coding mode according to the number of the pixels and the number of the data;
the second option is: interacting pixel coding parameters with the terminal, wherein the pixel coding parameters at least comprise: pixel coding method.
In some embodiments, the pixel encoding parameters include the pixel encoding mode, and in other implementations, the pixel encoding parameters may further include:
combining values of N x M of the coding modes;
the number of bits occupied by one pixel coded data of a single coding mode and/or a combined coding mode;
and coding the mapping relation.
In some embodiments, the MEC server may also not receive the mapping coding parameters from terminals, for example, a terminal held by a main broadcast is commonly used for live broadcast, and live broadcast generates the second type of three-dimensional video data. The MEC server finds that the pixel coding data comes from a specific terminal, and adopts default pixel coding parameters to restore the pixel values. The default pixel encoding parameters herein may include: default coding modes and/or pixel coding relationships.
In some embodiments, the pixel decoding the pixel encoding data according to the pixel encoding manner to obtain the pixel value of the second type of three-dimensional video data includes at least one of:
decoding the pixel coding data of a single pixel to restore a second type of three-dimensional video data pixel value according to a single pixel coding mode;
and decoding the pixel coding data of the N x M pixels to restore the pixel value of the three-dimensional video data of the second type according to a combined pixel coding mode.
In some embodiments, the step 302 may include: and inquiring a pixel coding mapping relation according to the pixel coding data to obtain a pixel value corresponding to the pixel coding data.
In some embodiments, the method further comprises:
and receiving the updated pixel coding mapping relation or the updated part in the pixel coding mapping relation sent by the terminal before restoring the pixel value of the second type of three-dimensional video data according to the pixel coding data.
And through the interaction of the pixel coding mapping relation, the pixel coding mapping relation is synchronized in the terminal and the MEC server.
In some embodiments, the MEC server may perform three-dimensional video modeling directly based on the second type of three-dimensional video data, or may perform modeling using the first type of three-dimensional video data after the restored second type of three-dimensional video data is restored and further converted into the first type of three-dimensional video data.
In some embodiments, the method further comprises:
receiving a conversion identifier sent by a terminal; in some cases, if the terminal performs conversion of different types of three-dimensional video data, the terminal may notify the MEC server, and at this time, the MEC server may receive a conversion identifier;
and converting the restored three-dimensional data of the second type into three-dimensional video data of the first type according to the conversion identifier.
In some cases, the MEC server is not notified by the terminal even if the conversion of the different types of three-dimensional video data is performed, and the MEC server performs three-dimensional video modeling directly based on the second type of three-dimensional video data.
In some further embodiments, the MEC server receives a terminal identifier sent by a terminal, and if the three-dimensional video modeling is successfully performed according to the second type of three-dimensional video data, the MEC server directly performs the three-dimensional video modeling by using the second type of three-dimensional video data, otherwise attempts to convert the second type of three-dimensional video data into the first type of three-dimensional video data; if the modeling of the three-dimensional video data converted into the first type is successful, continuing to convert the three-dimensional video data of the second type into the three-dimensional video data of the first type, and continuing to perform three-dimensional video modeling by using the three-dimensional video data of the first type; otherwise the transmission may be considered erroneous.
As shown in fig. 5, the present embodiment further provides a data processing apparatus, applied in a terminal, including:
a conversion module 400, configured to convert, if the acquired three-dimensional video data of the first type is obtained, the three-dimensional video data of the first type into three-dimensional video data of a second type;
the encoding module 401 is configured to perform pixel encoding according to the pixel value of the second type of three-dimensional video data to obtain pixel encoded data;
a sending module 402, configured to send the pixel encoding data to a moving edge computing MEC server, where the pixel encoding data is used for the MEC server to restore the second type of three-dimensional video data;
the data volume of the second type of three-dimensional video data before pixel coding is a first data volume; the data volume of the second type of three-dimensional video data after pixel coding is a second data volume; the first amount of data is greater than the second amount of data.
In some embodiments, the converting module 400, the encoding module 401 and the transmitting module 402 may be program modules corresponding to computer executable code, and the computer executable code is executed to enable the aforementioned transmission of the pixel encoding data and the second type of three-dimensional video data.
In other embodiments, the conversion module 400, the encoding module 401, and the sending module 402 may also be a combination of hardware modules and program modules, such as a complex programmable array or a field programmable array.
In still other embodiments, the conversion module 400, the encoding module 401, and the sending module 402 may correspond to hardware modules, for example, the conversion module 400, the encoding module 401, and the sending module 402 may be application specific integrated circuits.
In some embodiments, the encoding module 401 includes:
the color coding submodule is used for carrying out color coding according to the color pixel value of the second type of three-dimensional video data to obtain color coded data;
and/or the presence of a gas in the gas,
and the depth value coding submodule is used for coding the depth value according to the depth value pixel value of the second type of three-dimensional video data to obtain depth value coded data.
In some embodiments, the encoding module 401 is specifically configured to query a pixel encoding mapping relationship according to a pixel value of the second type of three-dimensional video data, and determine the pixel encoding data.
In some embodiments, the apparatus further comprises:
the selecting module is used for selecting a pixel coding mode according to preset information, wherein the preset information comprises: at least one of network transmission status information, load status information of the terminal, and load status information of the MEC server;
the encoding module 401 is specifically configured to perform pixel encoding on the pixel value according to a pixel encoding manner to obtain the pixel encoded data.
In some embodiments, the encoding module is at least to perform at least one of: according to a single-pixel coding mode, carrying out single-pixel coding on the pixel value of a single pixel of the second type of three-dimensional video data to obtain first type of coded data, wherein the number of bits occupied by the first type of coded data is smaller than the number of bits occupied by the pixel value; and performing combined pixel coding on the pixel values of N x M pixels of the three-dimensional video data of the second type according to a combined pixel coding mode to obtain a second type of pixel coding, wherein N and M are both positive integers.
In some embodiments, the N x M pixels are distributed adjacently;
and the N M pixels are distributed at intervals according to a preset interval mode.
In some embodiments, the encoding module 401 includes:
the query submodule is used for querying the pixel coding mapping relation according to the pixel value of the second type of three-dimensional video data;
and the determining submodule is used for determining the pixel coding data according to the pixel coding value corresponding to the pixel value if the pixel value is in the pixel coding mapping relation.
In some embodiments, the apparatus further comprises:
and the updating module is used for updating the pixel coding mapping relation according to the pixel value if the pixel value is not in the pixel coding mapping relation, and sending the updated pixel coding mapping relation or an updated part in the pixel coding mapping relation to the MEC server.
In some embodiments, the encoding module 401 is specifically configured to obtain the sequence number of the pixel value of the second type of three-dimensional video data according to the ordering of the pixel value of the second type of three-dimensional video data according to a preset ordering manner.
As shown in fig. 6, the present embodiment provides a data processing apparatus applied to a mobile edge computing MEC server, including:
a receiving module 501, configured to receive pixel encoded data sent by a terminal;
a restoring module 502, configured to restore pixel values of the second type of three-dimensional video data according to the pixel encoding data; the data volume of the second type of three-dimensional video data before pixel coding is a first data volume; the data volume of the second type of three-dimensional video data after pixel coding is a second data volume; the first amount of data is greater than the second amount of data.
In some embodiments, the receiving module 501 and the restoring module 502 may be program modules corresponding to computer executable codes, and the computer executable codes can be executed to implement the aforementioned transmission of the pixel encoding data and the second type of three-dimensional video data.
In other embodiments, the receiving module 501 and the restoring module 502 can also be a combination of hardware modules and program modules, such as a complex programmable array or a field programmable array.
In still other embodiments, the receiving module 501 and the restoring module 502 may correspond to hardware modules, for example, the encoding module and the sending module may be application specific integrated circuits.
In some embodiments, the restoration module 502 includes at least one of:
the color restoration sub-module is used for restoring the color pixel value of the second type of three-dimensional video data according to the color coding data of the pixel coding data;
and the depth value restoring submodule is used for restoring the depth value pixel value of the second type of three-dimensional video data according to the depth value encoding data of the pixel encoding data.
In some embodiments, the apparatus further comprises:
the determining module is used for determining a pixel coding mode of the pixel coding data;
the restoring module 502 is specifically configured to perform pixel decoding on the pixel encoding data according to the pixel encoding manner to obtain a pixel value of the second type of three-dimensional video data.
In some embodiments, the determining module comprises at least one of:
the first determining submodule is used for determining the number of pixels contained in the second type of three-dimensional video data and determining the number of data of the pixel coding data; determining the pixel coding mode according to the number of the pixels and the number of the data;
a second determining submodule, configured to interact with the terminal for pixel encoding parameters, where the pixel encoding parameters at least include: pixel coding method.
In some embodiments, the restoring module 502 is specifically configured to perform at least one of:
decoding the pixel coding data of a single pixel to restore a second type of three-dimensional video data pixel value according to a single pixel coding mode;
and decoding the pixel coding data of the N x M pixels to restore the pixel value of the three-dimensional video data of the second type according to a combined pixel coding mode.
In some embodiments, the restoring module 502 is specifically configured to query a pixel coding mapping relationship according to the pixel coding data, and obtain a pixel value corresponding to the pixel coding data.
In some embodiments, the apparatus further comprises:
a receiving module 501, configured to receive an updated pixel coding mapping relationship or an updated portion in the pixel coding mapping relationship sent by the terminal before restoring the pixel value of the second type of three-dimensional video data according to the pixel coding data.
The present embodiments provide a computer storage medium having stored thereon computer instructions which, when executed by a processor, implement the steps of a data processing method applied in a terminal or an MEC server, for example, as one or more of the methods shown in fig. 2 to 4.
As shown in fig. 7, the embodiment provides an electronic device, which includes a memory, a processor, and computer instructions stored in the memory and executable on the processor, wherein the processor executes the instructions to implement the steps of the data processing method applied in the terminal or the MEC server, for example, one or more of the methods shown in fig. 2 to fig. 4 may be executed.
In some embodiments, the electronic device further comprises: and the communication interface can be used for interacting information with other equipment. For example, if the electronic device is a terminal, the communication interface may perform information interaction at least with the MEC server. If the electronic equipment is an MEC server, the communication interface can at least perform information interaction with the terminal.
One specific example is provided below in connection with any of the embodiments described above:
numbering common colors in sequence, after the mobile phone collects two-dimensional (2D) video data in a YUV format, scanning a YUV value of each pixel of an image, and obtaining an RGB value corresponding to each pixel in the image through a conversion formula of the YUV value and the RGB value; if the RGB value is in the color sequence, replacing the RGB value data with the color sequence number; specifically, the RGB values corresponding to all pixels of the whole image are obtained, then the RGB value corresponding to each pixel is replaced by a serial number based on the numbering performed on the colors in advance, and then the pixels and the corresponding color serial numbers are packed and uploaded.
Numbering common colors in sequence, after the mobile phone collects red, green and blue (RGB) data, scanning the RGB data of each pixel of the image, and if the RGB data are in a color sequence, replacing the RGB data with color sequence numbers; specifically, RGB data of each pixel of an image are scanned, RGB data of the whole image pixel are counted, then the RGB is sequenced and numbered, the RGB of each pixel is replaced by a serial number, and then the pixel and the counted RGB data are packaged and uploaded; the MEC server and the mobile phone end store a mapping table, when RGB data are transmitted, pixels are scanned horizontally, if the pixels are not in the mapping table, a mapping (such as pixel RGB-mark A [16 bit ] or [32 bit ] or [8 bit ]) is created and stored in the mapping table, meanwhile, the RGB data are replaced by 16-bit color serial numbers, and after scanning is finished, changed items in the mapping table and the RGB data are uploaded, or the coding of a single pixel can be extended to NxN pixels to be coded together.
In the embodiments provided in the present invention, it should be understood that the disclosed method and intelligent device may be implemented in other ways. The above-described device embodiments are merely illustrative, for example, the division of the unit is only a logical functional division, and there may be other division ways in actual implementation, such as: multiple units or components may be combined, or may be integrated into another system, or some features may be omitted, or not implemented. In addition, the coupling, direct coupling or communication connection between the components shown or discussed may be through some interfaces, and the indirect coupling or communication connection between the devices or units may be electrical, mechanical or other forms.
The units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed on a plurality of network units; some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment.
In addition, all the functional units in the embodiments of the present invention may be integrated into one second processing unit, or each unit may be separately regarded as one unit, or two or more units may be integrated into one unit; the integrated unit can be realized in a form of hardware, or in a form of hardware plus a software functional unit.
Those of ordinary skill in the art will understand that: all or part of the steps for implementing the method embodiments may be implemented by hardware related to program instructions, and the program may be stored in a computer readable storage medium, and when executed, the program performs the steps including the method embodiments; and the aforementioned storage medium includes: a removable storage device, a ROM, a RAM, a magnetic or optical disk, or various other media that can store program code.
Alternatively, the integrated unit of the present invention may be stored in a computer-readable storage medium if it is implemented in the form of a software functional module and sold or used as a separate product. Based on such understanding, the technical solutions of the embodiments of the present invention may be essentially or partially implemented in the form of a software product, which is stored in a storage medium and includes several instructions to enable a computer device (which may be a personal computer, an MEC server, or a network device) to execute all or part of the methods described in the embodiments of the present invention. And the aforementioned storage medium includes: a removable storage device, a ROM, a RAM, a magnetic or optical disk, or various other media that can store program code.
It should be noted that: the technical schemes described in the embodiments of the present invention can be combined arbitrarily without conflict.
The above description is only for the specific embodiments of the present invention, but the scope of the present invention is not limited thereto, and any person skilled in the art can easily conceive of the changes or substitutions within the technical scope of the present invention, and all the changes or substitutions should be covered within the scope of the present invention.

Claims (32)

1. A data processing method is applied to a terminal and comprises the following steps:
if the three-dimensional video data of the first type is obtained, converting the three-dimensional video data of the first type into three-dimensional video data of a second type;
performing pixel coding according to the pixel value of the second type of three-dimensional video data to obtain pixel coding data;
sending the pixel encoding data to a Mobile Edge Computing (MEC) server, wherein the pixel encoding data is used for restoring the three-dimensional video data by the MEC server;
the data volume of the second type of three-dimensional video data before pixel coding is a first data volume; the data volume of the three-dimensional video data after pixel coding is a second data volume; the first amount of data is greater than the second amount of data;
the pixel encoding according to the pixel value of the second type of three-dimensional video data to obtain pixel encoded data includes:
inquiring a pixel coding mapping relation according to the pixel value of the second type of three-dimensional video data, and determining the pixel coding data; the pixel values include at least one of: color pixel values, depth value pixel values.
2. The method of claim 1,
the pixel encoding according to the pixel value of the second type of three-dimensional video data to obtain pixel encoded data includes:
performing color coding according to the color pixel value of the second type of three-dimensional video data to obtain color coded data;
and/or the presence of a gas in the gas,
and carrying out depth value encoding according to the depth value pixel value of the second type of three-dimensional video data to obtain depth value encoded data.
3. The method of claim 1,
the method further comprises the following steps:
selecting a pixel coding mode according to preset information, wherein the preset information comprises: at least one of network transmission status information, load status information of the terminal, and load status information of the MEC server;
the pixel encoding according to the pixel value of the second type of three-dimensional video data to obtain pixel encoded data includes:
and according to the selected pixel coding mode, carrying out pixel coding on the pixel value to obtain the pixel coding data.
4. The method of claim 3,
the pixel coding data obtained by pixel coding the pixel value according to the selected pixel coding mode comprises at least one of the following:
according to a single-pixel coding mode, carrying out single-pixel coding on the pixel value of a single pixel of the second type of three-dimensional video data to obtain first type of coded data, wherein the number of bits occupied by the first type of coded data is smaller than the number of bits occupied by the pixel value;
and performing combined pixel coding on the pixel values of N x M pixels of the three-dimensional video data of the second type according to a combined pixel coding mode to obtain a second type of pixel coding, wherein N and M are both positive integers.
5. The method of claim 4,
the N x M pixels are distributed adjacently;
and the N M pixels are distributed at intervals according to a preset interval mode.
6. The method of claim 1,
the querying a pixel coding mapping relation according to the pixel value of the second type of three-dimensional video data to determine the pixel coding data includes:
inquiring the pixel coding mapping relation according to the pixel value of the second type of three-dimensional video data;
and if the pixel value is in the pixel coding mapping relation, determining the pixel coding data according to the pixel coding value corresponding to the pixel value.
7. The method of claim 4, further comprising:
and if the pixel value is not in the pixel coding mapping relation, updating the pixel coding mapping relation according to the pixel value, and sending the updated pixel coding mapping relation or the updated part in the pixel coding mapping relation to the MEC server.
8. The method of claim 1,
the pixel encoding according to the pixel value of the second type of three-dimensional video data to obtain pixel encoded data includes:
and obtaining the pixel value sequence number of the second type of three-dimensional video data according to the sequence of the pixel values of the second type of three-dimensional video data in a preset sequence mode.
9. A data processing method is applied to a Mobile Edge Computing (MEC) server and comprises the following steps:
receiving pixel coded data sent by a terminal;
restoring pixel values of the second type of three-dimensional video data according to the pixel coding data; the data volume of the second type of three-dimensional video data before pixel coding is a first data volume; the data volume of the second type of three-dimensional video data after pixel coding is a second data volume; the first amount of data is greater than the second amount of data;
the restoring of the second type of three-dimensional video data from the pixel encoded data comprises:
inquiring pixel coding mapping relation according to pixel coding data of the second type of three-dimensional video data, and determining the pixel value; the pixel values include at least one of: color pixel values, depth value pixel values.
10. The method of claim 9,
restoring pixel values of the second type of three-dimensional video data according to the pixel encoding data, wherein the restoring comprises at least one of the following steps:
restoring color pixel values of the second type of three-dimensional video data according to color encoding data of the pixel encoding data;
and restoring the depth value pixel value of the second type of three-dimensional video data according to the depth value encoding data of the pixel encoding data.
11. The method according to claim 9 or 10,
the method further comprises the following steps:
determining a pixel coding mode of the pixel coding data;
the restoring of the second type of three-dimensional video data from the pixel encoded data comprises:
and according to the pixel coding mode, carrying out pixel decoding on the pixel coding data to obtain a pixel value of the second type of three-dimensional video data.
12. The method of claim 11,
the determining of the pixel coding mode of the pixel coding comprises at least one of the following steps:
determining the number of pixels contained in the second type of three-dimensional video data, and determining the number of data of the pixel coding data; determining the pixel coding mode according to the number of the pixels and the number of the data;
interacting pixel coding parameters with the terminal, wherein the pixel coding parameters at least comprise: pixel coding method.
13. The method of claim 11,
the pixel decoding of the pixel coded data according to the pixel coding mode to obtain the pixel value of the second type of three-dimensional video data includes at least one of the following:
decoding the pixel coding data of a single pixel to restore a second type of three-dimensional video data pixel value according to a single pixel coding mode;
and decoding the pixel coding data of the N x M pixels to restore the pixel value of the three-dimensional video data of the second type according to a combined pixel coding mode.
14. The method according to claim 9 or 10,
restoring pixel values of a second type of three-dimensional video data according to the pixel encoding data, comprising:
and inquiring a pixel coding mapping relation according to the pixel coding data to obtain a pixel value corresponding to the pixel coding data.
15. The method of claim 14, further comprising:
and receiving the updated pixel coding mapping relation or the updated part in the pixel coding mapping relation sent by the terminal before restoring the pixel value of the second type of three-dimensional video data according to the pixel coding data.
16. A data processing apparatus, applied to a terminal, comprising:
the conversion module is used for converting the three-dimensional video data of the first type into three-dimensional video data of a second type if the three-dimensional video data of the first type is acquired;
the encoding module is used for carrying out pixel encoding according to the pixel values of the second type of three-dimensional video data to obtain pixel encoding data;
a sending module, configured to send the pixel encoding data to a moving edge computation MEC server, where the pixel encoding data is used for the MEC server to restore the second type of three-dimensional video data;
the data volume of the second type of three-dimensional video data before pixel coding is a first data volume; the data volume of the second type of three-dimensional video data after pixel coding is a second data volume; the first amount of data is greater than the second amount of data;
the encoding module is specifically configured to query a pixel encoding mapping relationship according to the pixel value of the second type of three-dimensional video data, and determine the pixel encoding data; the pixel values include at least one of: color pixel values, depth value pixel values.
17. The apparatus of claim 16,
the encoding module includes:
the color coding submodule is used for carrying out color coding according to the color pixel value of the second type of three-dimensional video data to obtain color coded data;
and/or the presence of a gas in the gas,
and the depth value coding submodule is used for coding the depth value according to the depth value pixel value of the second type of three-dimensional video data to obtain depth value coded data.
18. The apparatus of claim 16, further comprising:
the selecting module is used for selecting a pixel coding mode according to preset information, wherein the preset information comprises: at least one of network transmission status information, load status information of the terminal, and load status information of the MEC server;
the encoding module is specifically configured to perform pixel encoding on the pixel values according to a pixel encoding mode to obtain the pixel encoded data.
19. The apparatus of claim 18,
the encoding module is at least configured to perform at least one of: according to a single-pixel coding mode, carrying out single-pixel coding on the pixel value of a single pixel of the second type of three-dimensional video data to obtain first type of coded data, wherein the number of bits occupied by the first type of coded data is smaller than the number of bits occupied by the pixel value; and performing combined pixel coding on the pixel values of N x M pixels of the three-dimensional video data of the second type according to a combined pixel coding mode to obtain a second type of pixel coding, wherein N and M are both positive integers.
20. The apparatus of claim 19,
the N x M pixels are distributed adjacently;
and the N M pixels are distributed at intervals according to a preset interval mode.
21. The apparatus of claim 16,
the encoding module includes:
the query submodule is used for querying the pixel coding mapping relation according to the pixel value of the second type of three-dimensional video data;
and the determining submodule is used for determining the pixel coding data according to the pixel coding value corresponding to the pixel value if the pixel value is in the pixel coding mapping relation.
22. The apparatus of claim 21, further comprising:
and the updating module is used for updating the pixel coding mapping relation according to the pixel value if the pixel value is not in the pixel coding mapping relation, and sending the updated pixel coding mapping relation or an updated part in the pixel coding mapping relation to the MEC server.
23. The apparatus of claim 16,
the encoding module is specifically configured to obtain a sequence number of pixel values of the second type of three-dimensional video data according to a sequence of the pixel values of the second type of three-dimensional video data according to a preset sequencing manner.
24. A data processing apparatus, applied to a mobile edge computing MEC server, comprising:
the receiving module is used for receiving pixel coding data sent by a terminal;
the restoration module is used for restoring the pixel values of the second type of three-dimensional video data according to the pixel coding data; the data volume of the second type of three-dimensional video data before pixel coding is a first data volume; the data volume of the second type of three-dimensional video data after pixel coding is a second data volume; the first amount of data is greater than the second amount of data.
25. The apparatus of claim 24,
the reduction module comprises at least one of the following:
the color restoration sub-module is used for restoring the color pixel value of the second type of three-dimensional video data according to the color coding data of the pixel coding data;
the depth value restoring submodule is used for restoring the depth value pixel value of the second type of three-dimensional video data according to the depth value encoding data of the pixel encoding data;
the restoring of the second type of three-dimensional video data from the pixel encoded data comprises:
inquiring pixel coding mapping relation according to pixel coding data of the second type of three-dimensional video data, and determining the pixel value; the pixel values include at least one of: color pixel values, depth value pixel values.
26. The apparatus of claim 24 or 25,
the device further comprises:
and the determining module is used for determining the pixel coding mode of the pixel coding data.
27. The apparatus of claim 26,
the determination module comprises at least one of:
the first determining submodule is used for determining the number of pixels contained in the second type of three-dimensional video data and determining the number of data of the pixel coding data; determining the pixel coding mode according to the number of the pixels and the number of the data;
a second determining submodule, configured to interact with the terminal for pixel encoding parameters, where the pixel encoding parameters at least include: pixel coding method.
28. The apparatus of claim 26,
the restoring module is specifically configured to execute at least one of:
decoding the pixel coding data of a single pixel to restore a second type of three-dimensional video data pixel value according to a single pixel coding mode;
and decoding the pixel coding data of the N x M pixels to restore the pixel value of the three-dimensional video data of the second type according to a combined pixel coding mode.
29. The apparatus of claim 24 or 25,
the restoring module is specifically configured to query a pixel coding mapping relationship according to the pixel coding data, and obtain a pixel value corresponding to the pixel coding data.
30. The apparatus of claim 29, further comprising:
and the receiving module is used for receiving the updated pixel coding mapping relation or the updated part in the pixel coding mapping relation sent by the terminal before restoring the pixel value of the second type of three-dimensional video data according to the pixel coding data.
31. A computer storage medium having computer instructions stored thereon, which when executed by a processor, perform the steps of the data processing method of any one of claims 1 to 8; alternatively, the instructions when executed by the processor implement the steps of the data processing method of any of claims 9 to 15.
32. An electronic device comprising a memory, a processor and computer instructions stored on the memory and executable on the processor, wherein the processor implements the steps of the data processing method of any one of claims 1 to 8 or implements the steps of the data processing method of any one of claims 9 to 15 when executing the instructions.
CN201811163098.7A 2018-09-30 2018-09-30 Data processing method and device, electronic equipment and storage medium Active CN109274976B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811163098.7A CN109274976B (en) 2018-09-30 2018-09-30 Data processing method and device, electronic equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811163098.7A CN109274976B (en) 2018-09-30 2018-09-30 Data processing method and device, electronic equipment and storage medium

Publications (2)

Publication Number Publication Date
CN109274976A CN109274976A (en) 2019-01-25
CN109274976B true CN109274976B (en) 2021-07-20

Family

ID=65195594

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811163098.7A Active CN109274976B (en) 2018-09-30 2018-09-30 Data processing method and device, electronic equipment and storage medium

Country Status (1)

Country Link
CN (1) CN109274976B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109257609B (en) * 2018-09-30 2021-04-23 Oppo广东移动通信有限公司 Data processing method and device, electronic equipment and storage medium
CN112672132B (en) 2018-09-30 2023-12-26 Oppo广东移动通信有限公司 Data processing method and device, electronic equipment and storage medium
CN111225219B (en) * 2019-11-26 2021-12-14 深圳英伦科技股份有限公司 Light field three-dimensional immersion experience information transmission method and system based on 5G network
CN111385579B (en) * 2020-04-09 2022-07-26 广州市百果园信息技术有限公司 Video compression method, device, equipment and storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105611274A (en) * 2016-01-08 2016-05-25 湖南拓视觉信息技术有限公司 Method and device for transmitting three-dimensional image data and three-dimensional imaging system
CN108495112A (en) * 2018-05-10 2018-09-04 Oppo广东移动通信有限公司 Data transmission method and terminal, computer storage media
CN108600168A (en) * 2018-03-19 2018-09-28 网御安全技术(深圳)有限公司 A kind of secure coding method and system for the attack of artificial intelligence image identification

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140009576A1 (en) * 2012-07-05 2014-01-09 Alcatel-Lucent Usa Inc. Method and apparatus for compressing, encoding and streaming graphics

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105611274A (en) * 2016-01-08 2016-05-25 湖南拓视觉信息技术有限公司 Method and device for transmitting three-dimensional image data and three-dimensional imaging system
CN108600168A (en) * 2018-03-19 2018-09-28 网御安全技术(深圳)有限公司 A kind of secure coding method and system for the attack of artificial intelligence image identification
CN108495112A (en) * 2018-05-10 2018-09-04 Oppo广东移动通信有限公司 Data transmission method and terminal, computer storage media

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
Augmented reality based on edge computing using the example of remote live support;Michael Schneider 等;《2017 IEEE International Conference on Industrial Technology (ICIT)》;20170504;正文第III-V节 *
delay constrained energy optimization for edge cloud offloading;Shreya Tayade 等;《2018 IEEE International Conference on Communications Workshops (ICC Workshops)》;20180705;全文 *
Michael Schneider 等.Augmented reality based on edge computing using the example of remote live support.《2017 IEEE International Conference on Industrial Technology (ICIT)》.2017, *

Also Published As

Publication number Publication date
CN109274976A (en) 2019-01-25

Similar Documents

Publication Publication Date Title
CN109151436B (en) Data processing method and device, electronic equipment and storage medium
CN109274976B (en) Data processing method and device, electronic equipment and storage medium
CN109257609B (en) Data processing method and device, electronic equipment and storage medium
CN103281539B (en) Method, device and the terminal that a kind of image coding and decoding processes
CN109194946B (en) Data processing method and device, electronic equipment and storage medium
US10848787B2 (en) Lossy image compression using palettization of locally mixed colors
CN112116551A (en) Camera shielding detection method and device, electronic equipment and storage medium
CN105282550A (en) Encoding method and decoding method, encoding and decoding system, encoder and decoder
CN110720223B (en) Method for video coding using dual deblocking filtering thresholds
CN111246208B (en) Video processing method and device and electronic equipment
WO2020063171A1 (en) Data transmission method, terminal, server and storage medium
CN109389674B (en) Data processing method and device, MEC server and storage medium
CN109309839B (en) Data processing method and device, electronic equipment and storage medium
RU2799771C2 (en) Methods and devices for data processing, electronic device
CN109389675B (en) Data processing method and device, terminal and storage medium
CN112087636B (en) Image coding processing method and device, storage medium and terminal
US20240020885A1 (en) Point cloud data transmission method, point cloud data transmission device, point cloud data reception method, and point cloud data reception device
US20230111994A1 (en) Point cloud data transmission device, point cloud data transmission method, point cloud data reception device, and point cloud data reception method
CN118301357A (en) Pixel chromaticity prediction method and device, electronic equipment and storage medium
Ge et al. E-LETTER

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant