WO2023241289A1 - Method and device for generating virtual reality service video, and storage medium - Google Patents

Method and device for generating virtual reality service video, and storage medium Download PDF

Info

Publication number
WO2023241289A1
WO2023241289A1 PCT/CN2023/094580 CN2023094580W WO2023241289A1 WO 2023241289 A1 WO2023241289 A1 WO 2023241289A1 CN 2023094580 W CN2023094580 W CN 2023094580W WO 2023241289 A1 WO2023241289 A1 WO 2023241289A1
Authority
WO
WIPO (PCT)
Prior art keywords
video
service
customer service
point cloud
cloud data
Prior art date
Application number
PCT/CN2023/094580
Other languages
French (fr)
Chinese (zh)
Inventor
甘仔斌
Original Assignee
中兴通讯股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 中兴通讯股份有限公司 filed Critical 中兴通讯股份有限公司
Publication of WO2023241289A1 publication Critical patent/WO2023241289A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/46Extracting features or characteristics from the video content, e.g. video fingerprints, representative shots or key frames
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/762Arrangements for image or video recognition or understanding using pattern recognition or machine learning using clustering, e.g. of similar faces in social networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/64Three-dimensional objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification

Definitions

  • Embodiments of the present disclosure relate to the field of communications, and specifically, to a method, device, and storage medium for generating virtual reality service videos.
  • the contact center in each system aims to allocate professionally trained customer service personnel to provide users with corresponding services, such as business inquiries, new business processing, reporting and complaints, and other services.
  • Customer service staff usually provide corresponding services to users through audio and video.
  • this service method like offline counters, one customer service staff can only provide services to one user at the same time, and there is still the problem of low service efficiency.
  • VR virtual reality
  • 5G networks With the rapid development of virtual reality (VR) technology and 5G networks, more and more VR technologies have been applied in 5G networks. If VR technology is applied to customer service, service efficiency can be improved through VR customer service. In order to make VR customer service services more realistic, it is necessary to convert VR virtual videos based on videos of customer service personnel providing services to users.
  • Embodiments of the present disclosure provide a method, device, and storage medium for generating virtual reality service videos to at least solve the problem of low service efficiency of online video services in related technologies.
  • a method for generating a virtual reality service video including: extracting 3D point cloud data in the service video, wherein the above-mentioned service video is a video formed by a customer service object performing customer service service; according to the above-mentioned service According to the optical flow parameters of the video, the above-mentioned customer service object is determined from each object included in the above-mentioned service video; the point cloud data of the above-mentioned customer service object is extracted from the above-mentioned 3D point cloud data, and the object pose parameters of the above-mentioned customer service object are determined; according to The above-mentioned object pose parameters convert the reference object model, and the point cloud data of the converted reference object model is brought into the above-mentioned 3D point cloud data to form a reference video, where the above-mentioned reference object model is the virtual reality customer service corresponding to the above-mentioned customer service object. ; Adjust the above-mentioned reference
  • a device for generating a virtual reality service video including: an extraction module configured to extract 3D point cloud data in the service video, wherein the above-mentioned service video is formed for a customer service object performing customer service.
  • the determination module is configured to determine the above-mentioned customer service object from each object included in the above-mentioned service video according to the optical flow parameters of the above-mentioned service video;
  • the pose module is configured to extract the above-mentioned customer service object from the above-mentioned 3D point cloud data point cloud data, and determine the object pose parameters of the above-mentioned customer service object;
  • the conversion module is configured to convert the reference object model according to the above-mentioned object pose parameters, and bring the point cloud data of the converted reference object model into the above-mentioned 3D
  • the point cloud data forms a reference video, wherein the reference object model is the virtual reality customer service corresponding to the customer service object; the adjustment module is configured to adjust the reference video until the optical flow parameters of the reference video meet the preset conditions, and obtain the same as the customer service object. Corresponding virtual reality service video.
  • a computer-readable storage medium is also provided, the computer-readable storage medium
  • a computer program is stored in the medium, wherein the computer program is configured to execute the steps in any of the above method embodiments when running.
  • an electronic device including a memory and a processor.
  • a computer program is stored in the memory, and the processor is configured to run the computer program to perform any of the above. Steps in method embodiments.
  • Figure 1 is a hardware structure block diagram of a method for generating virtual reality service videos according to an embodiment of the present disclosure
  • Figure 2 is a flow chart of a method for generating a virtual reality service video according to an embodiment of the present disclosure
  • Figure 3 is a flow chart of a method for generating a virtual reality service video according to an embodiment of the present disclosure
  • Figure 4 is a structural block diagram of a device for generating a virtual reality service video according to an embodiment of the present disclosure.
  • VR Virtual Reality, virtual reality technology
  • GAN Generative Adversarial Networks, Generative Adversarial Networks, a generative technology in deep learning
  • SLAM Simultaneous Localization and Mapping, simultaneous positioning and mapping technology, a method used for mapping and positioning in the field of robotics;
  • 3D pose the position, speed and acceleration of an object in three-dimensional space
  • 3D model the components and attributes of objects in virtual reality, including three-dimensional size, color, etc.;
  • Video service a video used to guide users to understand a certain business or guide the business process
  • VR service dynamically provide users with business processing services through characters in a virtual reality environment.
  • FIG. 1 is a hardware structure block diagram of a mobile terminal of a method for generating a virtual reality service video according to an embodiment of the present disclosure.
  • the mobile terminal may include one or more (only one is shown in Figure 1) processors 102 (the processor 102 may include but is not limited to a microprocessor (Central Processing Unit, MCU) or a programmable logic device (Field Programmable Gate Array, FPGA) and other processing devices) and a memory 104 for storing data, wherein the above-mentioned mobile terminal may also include a transmission device 106 for communication functions and an input and output device 108.
  • processors 102 may include but is not limited to a microprocessor (Central Processing Unit, MCU) or a programmable logic device (Field Programmable Gate Array, FPGA) and other processing devices) and a memory 104 for storing data
  • the above-mentioned mobile terminal may also include a transmission device 106 for communication functions and an input and output device 108.
  • the structure shown in Figure 1 is only illustrative, and it does not limit the structure of the above-mentioned mobile terminal.
  • the mobile terminal may also include more or fewer components than
  • the memory 104 can be used to store computer programs, for example, software programs and modules of application software, such as the computer program corresponding to the method for generating a virtual reality service video in the embodiment of the present disclosure.
  • the processor 102 runs the computer program stored in the memory 104 , thereby executing various functional applications and data processing, that is, implementing the above method.
  • Memory 104 may include high-speed random access memory, and may also include non-volatile memory, such as one or more magnetic storage devices, flash memory, memory, or other non-volatile solid-state memory.
  • the memory 104 may further include memory located remotely relative to the processor 102, and these remote memories may be connected to the mobile terminal through a network. Examples of the above-mentioned networks include but are not limited to the Internet, intranets, local area networks, mobile communication networks and combinations thereof.
  • the transmission device 106 is used to receive or send data via a network.
  • Specific examples of the above-mentioned network may include a wireless network provided by a communication provider of the mobile terminal.
  • the transmission device 106 includes a network adapter (Network Interface Controller, NIC for short), which can be connected to other network devices through a base station to communicate with the Internet.
  • the transmission device 106 may be a radio frequency (Radio Frequency, RF for short) module, which is used to communicate with the Internet wirelessly.
  • NIC Network Interface Controller
  • Figure 2 is a flow chart according to an embodiment of the present disclosure. As shown in Figure 2, the process includes the following steps:
  • Step S202 extract the 3D point cloud data in the service video, where the service video is a video formed by the customer service object performing customer service service;
  • Step S204 determine the customer service object from each object included in the service video according to the optical flow parameters of the service video
  • Step S206 extract the point cloud data of the customer service object from the 3D point cloud data, and determine the object pose parameters of the customer service object;
  • Step S208 Convert the reference object model according to the object pose parameters, and bring the point cloud data of the converted reference object model into the 3D point cloud data to form a reference video, where the reference object model is the virtual reality customer service corresponding to the customer service object. ;
  • Step S210 adjust the reference video until the optical flow parameters of the reference video meet the preset conditions, and obtain the virtual reality service video corresponding to the customer service object.
  • Service videos include dynamic images of customer service objects performing customer service services, and are not limited to including image information such as the environment where the customer service objects are located.
  • the specific method of extracting the 3D point cloud data in the service video is not limited here.
  • the 3D point cloud data in the service video is extracted through SLAM technology.
  • the optical flow parameters in the service video are not limited to the optical flow continuity determined for adjacent key frames in the service video. According to the optical flow continuity, each included object is determined from the 3D point cloud data of the service video, and then each object is Customer service objects are identified.
  • the point cloud data of the customer service object is extracted from the 3D point cloud data of the service video, and the object pose parameter of the customer service object is obtained based on the point cloud data of the customer service object.
  • the object pose parameter indicates It is the position and posture of the customer service object, such as gestures and the degree of bending of joints. For each key frame, it is not limited to obtaining the corresponding pose parameters.
  • the reference object model is transformed into a pose, so that the point cloud data converted to the reference object model corresponding to the pose is brought into the 3D point cloud data after the customer service object is extracted from the service video. to form a reference video.
  • the reference object model is the character model of the virtual reality customer service corresponding to the customer service object, and is not limited to the character model having the characteristics of the customer service object and the reference image.
  • the adjustment of the reference video including the point cloud data of the reference object model is not limited to the comparison of the optical flow parameters of the reference video and the optical flow parameters of the service video.
  • the pose of the reference object model is adjusted to obtain the optical flow parameters.
  • Virtual reality service videos that meet preset conditions.
  • the reference object model is converted through the pose parameters of the customer service object in the service video, and the converted reference object model is brought into the 3D point cloud data of the service video to form a reference video, so that the reference view is
  • the optical flow parameters of the frequency are adjusted to obtain the virtual reality service video.
  • the corresponding virtual reality service video is generated based on the service video of the customer service object.
  • the virtual reality service video is used for customer service, breaking the problem that the same customer service object can only serve one user at the same time.
  • use virtual reality service videos to provide customer service services to multiple users at the same time. Therefore, the problem of low service efficiency of online video services can be solved, and the technical effect of improving the service efficiency of video services can be achieved.
  • the above further includes:
  • S12 use neural network to convert the facial data and reference image features of the customer service object into a reference object model.
  • the reference object model is not limited to the neural network, and the reference object model will be generated by overcoming the object's facial data and reference image features.
  • the facial data of the customer service object is not limited to the facial feature data of the customer service object
  • the reference image features are not limited to the preset body image features, including hairstyle, limbs, clothing, accessories and other character features other than the face.
  • the facial data of the customer service object and the preset reference image characteristics are used to generate a virtual reality (VR) customer service object corresponding to the customer service object, and the VR customer service object is used to further generate a VR service video, so that the VR service video can replace the customer service object.
  • VR virtual reality
  • the form of video service improves the service efficiency of video services.
  • the above-mentioned S204 determines the customer service object from each object included in the service video according to the optical flow parameters of the service video, including:
  • S204-2 Use the key frame images in the service video to identify each object to determine the customer service object from each object.
  • the above-mentioned clustering of 3D point cloud data is performed based on the optical flow parameters of the service video to obtain various objects included in the service video, including:
  • S204-13 utilize the continuity of optical flow parameters of adjacent key frames to obtain each object through clustering.
  • the above method uses key frame images in the service video to identify each object, including:
  • S204-23 Match the recognition results with each object, and add object labels to each object according to the matching results.
  • the clustering result is not limited to each 3D object included in the service video obtained by clustering.
  • the key frame image is segmented from the service video according to the clustering result. It is not limited to determining the object for each key frame image from the service video.
  • the identified keyframe image for example, the keyframe image that best matches the clustering result in any one or more dimensions such as the number of objects, shape, etc.
  • the object recognition results are not limited to identifying the specific classification of each object in the key frame image, such as people, microphones, exhibition boards, etc.
  • the above-mentioned S206 extracts the point cloud data of the customer service object from the 3D point cloud data, and determines the object pose parameters of the customer service object, including:
  • S206-2 Based on the point cloud data of the customer service object, calculate the object pose of the customer service object and the probability corresponding to the object pose.
  • the customer service objects included in the 3D objects are not limited to being determined through the object labels.
  • the 3D objects identified as "characters" are determined as customer service objects, and the 3D objects of the service videos are determined. Extract point cloud data of customer service objects from point cloud data.
  • the object pose of the customer service object When the point cloud data of the customer service object is extracted, the object pose of the customer service object and the probability corresponding to the object pose are calculated.
  • the object pose of the customer service object is not limited to the object pose of the customer service object corresponding to each key frame of the service video, and the probability change interval of the object pose.
  • the object pose is not limited to the 3D pose of the customer service object.
  • the above-mentioned S210 adjusts the reference video until the optical flow parameters of the reference video meet the preset conditions, and obtains the virtual reality service video corresponding to the customer service object, including:
  • the point cloud data of the reference object model is not limited to the point cloud data corresponding to each key frame.
  • the reference object model is brought in for each key frame to obtain all key frames. Includes reference videos for VR customer service.
  • the reference video is adjusted until the adjusted optical flow difference parameter of the reference video is less than the preset threshold, and the reference video is used as the VR service video.
  • the adjustment is not limited to adjusting the pose of the VR customer service in each reference key frame. By adjusting the pose of the VR customer service brought in, the difference between the optical flow of the formed reference key frame and the optical flow of the original key frame is reduced to the predetermined level. Set threshold.
  • the generation of virtual reality VR service videos is not limited to that shown in Figure 3.
  • SLAM technology is used to extract the 3D point cloud data of the video service.
  • the optical flow of each key frame of the video service is obtained, and the optical flow continuity of adjacent key frames is used to perform point cloud clustering to determine the point cloud data of the customer service included in the video service.
  • each customer service point cloud data is calculated.
  • the 3D pose of the customer service agent in key frames After converting the VR character model corresponding to the customer service according to the 3D pose, the transformed VR character model is used to replace the customer service and bring it into the 3D point cloud data to form a reference view.
  • the VR customer service in the reference video is iterated until the VR service is obtained.
  • the method according to the above embodiments can be implemented by means of software plus the necessary general hardware platform. Of course, it can also be implemented by hardware, but in many cases the former is Better implementation.
  • the technical solution of the present disclosure can be embodied in the form of a software product in nature or in part that contributes to the existing technology.
  • the computer software product is stored in a storage medium (such as read-only memory/random access memory).
  • the memory Read-Only Memory/Random Access Memory, ROM/RAM), magnetic disk, optical disk
  • includes several instructions to cause a terminal device (which can be a mobile phone, computer, server, or network device, etc.) to execute the disclosure Methods described in various embodiments.
  • This embodiment also provides a device for generating a virtual reality service video.
  • the device is configured to implement the above embodiments and preferred implementations. What has already been explained will not be described again.
  • the term "module” may be a combination of software and/or hardware that implements a predetermined function.
  • the apparatus described in the following embodiments is preferably implemented in software, implementation in hardware, or a combination of software and hardware, is also possible and contemplated.
  • Figure 4 is a structural block diagram of a device for generating virtual reality service videos according to an embodiment of the present disclosure. As shown in Figure 4, the device includes:
  • the extraction module 41 is configured to extract 3D point cloud data in the service video, where the service video is a video formed by the customer service object performing customer service service;
  • the determination module 42 is configured to determine the customer service object from each object included in the service video according to the optical flow parameters of the service video;
  • the pose module 43 is configured to extract the point cloud data of the customer service object from the 3D point cloud data, and determine the object pose parameters of the customer service object;
  • the conversion module 44 is configured to convert the reference object model according to the object pose parameters, and bring the point cloud data of the converted reference object model into the 3D point cloud data to form a reference video, where the reference object model is the corresponding one of the customer service object. virtual reality customer service;
  • the adjustment module 45 is configured to adjust the reference video until the optical flow parameters of the reference video meet the preset conditions to obtain a virtual reality service video corresponding to the customer service object.
  • the above-mentioned determination module 42 includes: clustering 3D point cloud data according to the optical flow parameters of the service video to obtain each object included in the service video; using key frame images in the service video to identify each object to Determine the customer service objects from each object.
  • the above-mentioned determination module 42 clusters the 3D point cloud data according to the optical flow parameters of the service video to obtain each object included in the service video, and further includes: determining the optical flow parameters of adjacent key frames in the service video; Compare the optical flow parameters with The two-dimensional coordinates formed by 3D point cloud data mapping are matched; the continuity of the optical flow parameters of adjacent key frames is used to cluster each object.
  • the above-mentioned determination module 42 uses the key frame images in the service video to identify each object, which also includes: segmenting the key frame images from the service video according to the clustering results; performing object recognition on the key frame images to obtain the identification Result; match the recognition results with each object, and add object labels to each object based on the matching results.
  • the above-mentioned pose module 43 also includes: extracting the point cloud data of the customer service object from the 3D point cloud data according to the object label; calculating the object pose and object pose of the customer service object based on the point cloud data of the customer service object. corresponding probability.
  • the above-mentioned generating device of virtual reality service video also includes a model module, which is configured to obtain the facial data of the customer service object before converting the reference object model according to the object pose parameters; and use the neural network to convert the facial data of the customer service object into and reference image features are transformed into reference object models.
  • a model module which is configured to obtain the facial data of the customer service object before converting the reference object model according to the object pose parameters; and use the neural network to convert the facial data of the customer service object into and reference image features are transformed into reference object models.
  • the above-mentioned adjustment module 45 includes: calculating the optical flow parameters of the reference key frames in the reference video; obtaining the optical flow parameters of the original key frames corresponding to the reference key frames in the service video; comparing the optical flow parameters of the reference key frames with the original key frames.
  • the optical flow parameter of the frame is obtained to obtain the optical flow difference parameter; when the optical flow difference parameter corresponding to each reference key frame of the reference video is less than the preset threshold, the virtual reality service video is determined to be obtained; when the optical flow difference parameter is greater than or When equal to the preset threshold, the pose of the reference object model is adjusted according to the optical flow difference parameter until the virtual reality service video is obtained.
  • the reference object model is converted through the pose parameters of the customer service object in the service video, and the converted reference object model is brought into the 3D point cloud data of the service video to form a reference video, so that the reference video is
  • the optical flow parameters are adjusted to obtain the virtual reality service video.
  • the corresponding virtual reality service video is generated based on the service video of the customer service object.
  • the virtual reality service video is used for customer service, breaking the problem that the same customer service object can only provide one user at the same time. Limitation of customer service, use virtual reality service video to provide customer service to multiple users at the same time. Therefore, the problem of low service efficiency of online video services can be solved, and the technical effect of improving the service efficiency of video services can be achieved.
  • each of the above modules can be implemented through software or hardware.
  • it can be implemented in the following ways, but is not limited to this: the above modules are all located in the same processor; or the above modules can be implemented in any combination.
  • the forms are located in different processors.
  • Embodiments of the present disclosure also provide a computer-readable storage medium that stores a computer program, wherein the computer program is configured to execute the steps in any of the above method embodiments when running.
  • the computer-readable storage medium may include but is not limited to: U disk, read-only memory (Read-Only Memory, referred to as ROM), random access memory (Random Access Memory, referred to as RAM) , mobile hard disk, magnetic disk or optical disk and other media that can store computer programs.
  • ROM read-only memory
  • RAM random access memory
  • mobile hard disk magnetic disk or optical disk and other media that can store computer programs.
  • Embodiments of the present disclosure also provide an electronic device, including a memory and a processor.
  • a computer program is stored in the memory, and the processor is configured to run the computer program to perform the steps in any of the above method embodiments.
  • the above-mentioned electronic device may further include a transmission device and an input-output device, wherein the transmission device is connected to the above-mentioned processor, and the input-output device is connected to the above-mentioned processor.
  • each module or each step of the above-mentioned embodiments of the present disclosure can be implemented by a general computing device, and they can be concentrated on a single computing device, or distributed among multiple computing devices. over a network, they may be implemented with program code executable by a computing device, such that they may be stored in a storage device for execution by the computing device, and in some cases, may be executed in a sequence different from that described here.
  • the steps shown or described may be implemented by fabricating them separately into individual integrated circuit modules, or by fabricating multiple modules or steps among them into a single integrated circuit module. As such, the present disclosure is not limited to any specific combination of hardware and software.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • General Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Artificial Intelligence (AREA)
  • Computing Systems (AREA)
  • Databases & Information Systems (AREA)
  • Evolutionary Computation (AREA)
  • Medical Informatics (AREA)
  • Software Systems (AREA)
  • Processing Or Creating Images (AREA)

Abstract

Embodiments of the present disclosure provide a method for generating a virtual reality service video, comprising: extracting 3D point cloud data in a service video, wherein the service video is a video formed by a customer service subject performing a customer service; determining, according to an optical flow parameter of the service video, the customer service subject from subjects comprised in the service video; extracting point cloud data of the customer service subject from the 3D point cloud data, and determining a subject pose parameter of the customer service subject; transforming a reference subject model according to the subject pose parameter, and substituting point cloud data of the transformed reference subject model into the 3D point cloud data to form a reference video, wherein the reference subject model is a virtual reality customer service corresponding to the customer service subject; and adjusting the reference video until an optical flow parameter of the reference video meets a preset condition, so as to obtain a virtual reality service video corresponding to the customer service subject.

Description

虚拟现实服务视频的生成方法、装置及存储介质Virtual reality service video generation method, device and storage medium 技术领域Technical field
本公开实施例涉及通信领域,具体而言,涉及一种虚拟现实服务视频的生成方法、装置及存储介质。Embodiments of the present disclosure relate to the field of communications, and specifically, to a method, device, and storage medium for generating virtual reality service videos.
背景技术Background technique
各个系统中的联络中心旨在分配经过专业训练的客服人员为用户提供相应服务,例如查询业务、新业务办理、举报投诉等服务。客服人员通常通过音视频为用户提供相应服务,但这种服务方式,与线下柜台一样,在同一时间段内一个客服人员只能为一个用户提供服务,依旧存在服务效率较低的问题。The contact center in each system aims to allocate professionally trained customer service personnel to provide users with corresponding services, such as business inquiries, new business processing, reporting and complaints, and other services. Customer service staff usually provide corresponding services to users through audio and video. However, in this service method, like offline counters, one customer service staff can only provide services to one user at the same time, and there is still the problem of low service efficiency.
随着虚拟现实(Virtual Reality,VR)技术与5G网络的高速发展,越来越多的VR技术在5G网络中实现了应用。如果将VR技术应用于客服服务中,能够通过VR客服服务提高服务效率。为使得VR客服服务更加真实,就需要基于客服人员在为用户提供服务的视频进行VR虚拟视频的转化。With the rapid development of virtual reality (VR) technology and 5G networks, more and more VR technologies have been applied in 5G networks. If VR technology is applied to customer service, service efficiency can be improved through VR customer service. In order to make VR customer service services more realistic, it is necessary to convert VR virtual videos based on videos of customer service personnel providing services to users.
发明内容Contents of the invention
本公开实施例提供了一种虚拟现实服务视频的生成方法、装置及存储介质,以至少解决相关技术中线上视频服务的服务效率较低的问题。Embodiments of the present disclosure provide a method, device, and storage medium for generating virtual reality service videos to at least solve the problem of low service efficiency of online video services in related technologies.
根据本公开的一个实施例,提供了一种虚拟现实服务视频的生成方法,包括:提取服务视频中的3D点云数据,其中,上述服务视频为客服对象执行客服服务形成的视频;根据上述服务视频的光流参数,从上述服务视频包括的各个对象中确定出上述客服对象;从上述3D点云数据中提取出上述客服对象的点云数据,并确定上述客服对象的对象位姿参数;根据上述对象位姿参数对参考对象模型进行转换,并将转换后的参考对象模型的点云数据带入上述3D点云数据形成参考视频,其中,上述参考对象模型为上述客服对象对应的虚拟现实客服;调整上述参考视频直至上述参考视频的光流参数满足预设条件,得到与上述客服对象对应的虚拟现实服务视频。According to an embodiment of the present disclosure, a method for generating a virtual reality service video is provided, including: extracting 3D point cloud data in the service video, wherein the above-mentioned service video is a video formed by a customer service object performing customer service service; according to the above-mentioned service According to the optical flow parameters of the video, the above-mentioned customer service object is determined from each object included in the above-mentioned service video; the point cloud data of the above-mentioned customer service object is extracted from the above-mentioned 3D point cloud data, and the object pose parameters of the above-mentioned customer service object are determined; according to The above-mentioned object pose parameters convert the reference object model, and the point cloud data of the converted reference object model is brought into the above-mentioned 3D point cloud data to form a reference video, where the above-mentioned reference object model is the virtual reality customer service corresponding to the above-mentioned customer service object. ; Adjust the above-mentioned reference video until the optical flow parameters of the above-mentioned reference video meet the preset conditions, and obtain the virtual reality service video corresponding to the above-mentioned customer service object.
根据本公开的另一个实施例,提供了一种虚拟现实服务视频的生成装置,包括:提取模块,设置为提取服务视频中的3D点云数据,其中,上述服务视频为客服对象执行客服服务形成的视频;确定模块,设置为根据上述服务视频的光流参数,从上述服务视频包括的各个对象中确定出上述客服对象;位姿模块,设置为从上述3D点云数据中提取出上述客服对象的点云数据,并确定上述客服对象的对象位姿参数;转换模块,设置为根据上述对象位姿参数对参考对象模型进行转换,并将转换后的参考对象模型的点云数据带入上述3D点云数据形成参考视频,其中,上述参考对象模型为上述客服对象对应的虚拟现实客服;调整模块,设置为调整上述参考视频直至上述参考视频的光流参数满足预设条件,得到与上述客服对象对应的虚拟现实服务视频。According to another embodiment of the present disclosure, a device for generating a virtual reality service video is provided, including: an extraction module configured to extract 3D point cloud data in the service video, wherein the above-mentioned service video is formed for a customer service object performing customer service. video; the determination module is configured to determine the above-mentioned customer service object from each object included in the above-mentioned service video according to the optical flow parameters of the above-mentioned service video; the pose module is configured to extract the above-mentioned customer service object from the above-mentioned 3D point cloud data point cloud data, and determine the object pose parameters of the above-mentioned customer service object; the conversion module is configured to convert the reference object model according to the above-mentioned object pose parameters, and bring the point cloud data of the converted reference object model into the above-mentioned 3D The point cloud data forms a reference video, wherein the reference object model is the virtual reality customer service corresponding to the customer service object; the adjustment module is configured to adjust the reference video until the optical flow parameters of the reference video meet the preset conditions, and obtain the same as the customer service object. Corresponding virtual reality service video.
根据本公开的又一个实施例,还提供了一种计算机可读存储介质,所述计算机可读存储 介质中存储有计算机程序,其中,所述计算机程序被设置为运行时执行上述任一项方法实施例中的步骤。According to yet another embodiment of the present disclosure, a computer-readable storage medium is also provided, the computer-readable storage medium A computer program is stored in the medium, wherein the computer program is configured to execute the steps in any of the above method embodiments when running.
根据本公开的又一个实施例,还提供了一种电子装置,包括存储器和处理器,所述存储器中存储有计算机程序,所述处理器被设置为运行所述计算机程序以执行上述任一项方法实施例中的步骤。According to yet another embodiment of the present disclosure, an electronic device is also provided, including a memory and a processor. A computer program is stored in the memory, and the processor is configured to run the computer program to perform any of the above. Steps in method embodiments.
附图说明Description of the drawings
图1是根据本公开实施例的虚拟现实服务视频的生成方法的硬件结构框图;Figure 1 is a hardware structure block diagram of a method for generating virtual reality service videos according to an embodiment of the present disclosure;
图2是根据本公开实施例的虚拟现实服务视频的生成方法的流程图;Figure 2 is a flow chart of a method for generating a virtual reality service video according to an embodiment of the present disclosure;
图3是根据本公开实施例的虚拟现实服务视频的生成方法的流程图;Figure 3 is a flow chart of a method for generating a virtual reality service video according to an embodiment of the present disclosure;
图4是根据本公开实施例的虚拟现实服务视频的生成装置的结构框图。Figure 4 is a structural block diagram of a device for generating a virtual reality service video according to an embodiment of the present disclosure.
具体实施方式Detailed ways
下文中将参考附图并结合实施例来详细说明本公开的实施例。Hereinafter, embodiments of the present disclosure will be described in detail with reference to the accompanying drawings and embodiments.
需要说明的是,本公开的说明书和权利要求书及上述附图中的术语“第一”、“第二”等是用于区别类似的对象,而不必用于描述特定的顺序或先后次序。It should be noted that the terms "first", "second", etc. in the description and claims of the present disclosure and the above-mentioned drawings are used to distinguish similar objects and are not necessarily used to describe a specific order or sequence.
术语说明:Terminology:
VR:Virtual Reality,虚拟现实技术;VR: Virtual Reality, virtual reality technology;
GAN:Generative Adversarial Networks,生成对抗网络,深度学习中的一种生成式的技术GAN: Generative Adversarial Networks, Generative Adversarial Networks, a generative technology in deep learning
SLAM:Simultaneous Localization and Mapping,同步定位与建图技术,机器人领域用于建图与定位的方法;SLAM: Simultaneous Localization and Mapping, simultaneous positioning and mapping technology, a method used for mapping and positioning in the field of robotics;
3D位姿:物体在三维空间中的位置,速度及加速度;3D pose: the position, speed and acceleration of an object in three-dimensional space;
3D模型:在虚拟现实中物体的组成模块与属性,包括三维大小,颜色等等;3D model: the components and attributes of objects in virtual reality, including three-dimensional size, color, etc.;
视频服务:用于指引用户认识某个业务,或者办理业务过程导引的视频;Video service: a video used to guide users to understand a certain business or guide the business process;
VR服务:在虚拟现实环境中通过人物,动态为用户提供业务办理的服务。VR service: dynamically provide users with business processing services through characters in a virtual reality environment.
本申请实施例中所提供的方法实施例可以在移动终端、计算机终端或者类似的运算装置中执行。以运行在移动终端上为例,图1是本公开实施例的一种虚拟现实服务视频的生成方法的移动终端的硬件结构框图。如图1所示,移动终端可以包括一个或多个(图1中仅示出一个)处理器102(处理器102可以包括但不限于微处理器(Central Processing Unit,MCU)或可编程逻辑器件(Field Programmable Gate Array,FPGA)等的处理装置)和用于存储数据的存储器104,其中,上述移动终端还可以包括用于通信功能的传输设备106以及输入输出设备108。本领域普通技术人员可以理解,图1所示的结构仅为示意,其并不对上述移动终端的结构造成限定。例如,移动终端还可包括比图1中所示更多或者更少的组件,或者具有与图1所示不同的配置。The method embodiments provided in the embodiments of this application can be executed in a mobile terminal, a computer terminal, or a similar computing device. Taking running on a mobile terminal as an example, FIG. 1 is a hardware structure block diagram of a mobile terminal of a method for generating a virtual reality service video according to an embodiment of the present disclosure. As shown in Figure 1, the mobile terminal may include one or more (only one is shown in Figure 1) processors 102 (the processor 102 may include but is not limited to a microprocessor (Central Processing Unit, MCU) or a programmable logic device (Field Programmable Gate Array, FPGA) and other processing devices) and a memory 104 for storing data, wherein the above-mentioned mobile terminal may also include a transmission device 106 for communication functions and an input and output device 108. Persons of ordinary skill in the art can understand that the structure shown in Figure 1 is only illustrative, and it does not limit the structure of the above-mentioned mobile terminal. For example, the mobile terminal may also include more or fewer components than shown in FIG. 1 , or have a different configuration than shown in FIG. 1 .
存储器104可用于存储计算机程序,例如,应用软件的软件程序以及模块,如本公开实施例中的虚拟现实服务视频的生成方法对应的计算机程序,处理器102通过运行存储在存储器104内的计算机程序,从而执行各种功能应用以及数据处理,即实现上述的方法。存储器104可包括高速随机存储器,还可包括非易失性存储器,如一个或者多个磁性存储装置、闪 存、或者其他非易失性固态存储器。在一些实例中,存储器104可进一步包括相对于处理器102远程设置的存储器,这些远程存储器可以通过网络连接至移动终端。上述网络的实例包括但不限于互联网、企业内部网、局域网、移动通信网及其组合。The memory 104 can be used to store computer programs, for example, software programs and modules of application software, such as the computer program corresponding to the method for generating a virtual reality service video in the embodiment of the present disclosure. The processor 102 runs the computer program stored in the memory 104 , thereby executing various functional applications and data processing, that is, implementing the above method. Memory 104 may include high-speed random access memory, and may also include non-volatile memory, such as one or more magnetic storage devices, flash memory, memory, or other non-volatile solid-state memory. In some examples, the memory 104 may further include memory located remotely relative to the processor 102, and these remote memories may be connected to the mobile terminal through a network. Examples of the above-mentioned networks include but are not limited to the Internet, intranets, local area networks, mobile communication networks and combinations thereof.
传输装置106用于经由一个网络接收或者发送数据。上述的网络具体实例可包括移动终端的通信供应商提供的无线网络。在一个实例中,传输装置106包括一个网络适配器(Network Interface Controller,简称为NIC),其可通过基站与其他网络设备相连从而可与互联网进行通讯。在一个实例中,传输装置106可以为射频(Radio Frequency,简称为RF)模块,其用于通过无线方式与互联网进行通讯。The transmission device 106 is used to receive or send data via a network. Specific examples of the above-mentioned network may include a wireless network provided by a communication provider of the mobile terminal. In one example, the transmission device 106 includes a network adapter (Network Interface Controller, NIC for short), which can be connected to other network devices through a base station to communicate with the Internet. In one example, the transmission device 106 may be a radio frequency (Radio Frequency, RF for short) module, which is used to communicate with the Internet wirelessly.
在本实施例中提供了一种虚拟现实服务视频的生成方法,图2是根据本公开实施例的的流程图,如图2所示,该流程包括如下步骤:In this embodiment, a method for generating a virtual reality service video is provided. Figure 2 is a flow chart according to an embodiment of the present disclosure. As shown in Figure 2, the process includes the following steps:
步骤S202,提取服务视频中的3D点云数据,其中,服务视频为客服对象执行客服服务形成的视频;Step S202, extract the 3D point cloud data in the service video, where the service video is a video formed by the customer service object performing customer service service;
步骤S204,根据服务视频的光流参数,从服务视频包括的各个对象中确定出客服对象;Step S204, determine the customer service object from each object included in the service video according to the optical flow parameters of the service video;
步骤S206,从3D点云数据中提取出客服对象的点云数据,并确定客服对象的对象位姿参数;Step S206, extract the point cloud data of the customer service object from the 3D point cloud data, and determine the object pose parameters of the customer service object;
步骤S208,根据对象位姿参数对参考对象模型进行转换,并将转换后的参考对象模型的点云数据带入3D点云数据形成参考视频,其中,参考对象模型为客服对象对应的虚拟现实客服;Step S208: Convert the reference object model according to the object pose parameters, and bring the point cloud data of the converted reference object model into the 3D point cloud data to form a reference video, where the reference object model is the virtual reality customer service corresponding to the customer service object. ;
步骤S210,调整参考视频直至参考视频的光流参数满足预设条件,得到与客服对象对应的虚拟现实服务视频。Step S210, adjust the reference video until the optical flow parameters of the reference video meet the preset conditions, and obtain the virtual reality service video corresponding to the customer service object.
服务视频包括客服对象执行客服服务的动态图像,还不限于包括客服对象所处环境等图像信息。提取服务视频中的3D点云数据的具体方式在此不作限定,例如通过SLAM技术提取服务视频中的3D点云数据。Service videos include dynamic images of customer service objects performing customer service services, and are not limited to including image information such as the environment where the customer service objects are located. The specific method of extracting the 3D point cloud data in the service video is not limited here. For example, the 3D point cloud data in the service video is extracted through SLAM technology.
服务视频中的光流参数不限于为服务视频中相邻关键帧确定出的光流连续性,根据光流连续性从服务视频的3D点云数据中确定出包括的各个对象,继而从各个对象中识别出客服对象。The optical flow parameters in the service video are not limited to the optical flow continuity determined for adjacent key frames in the service video. According to the optical flow continuity, each included object is determined from the 3D point cloud data of the service video, and then each object is Customer service objects are identified.
在识别出客服对象的情况下,从服务视频的3D点云数据中提取出客服对象的点云数据,并根据客服对象的点云数据得到客服对象的对象位姿参数,对象位姿参数指示的是客服对象的位置姿势,例如手势、关节的弯曲程度。针对每幅关键帧不限于获取对应的位姿参数。When the customer service object is identified, the point cloud data of the customer service object is extracted from the 3D point cloud data of the service video, and the object pose parameter of the customer service object is obtained based on the point cloud data of the customer service object. The object pose parameter indicates It is the position and posture of the customer service object, such as gestures and the degree of bending of joints. For each key frame, it is not limited to obtaining the corresponding pose parameters.
根据每幅关键帧的位姿参数,对参考对象模型进行位姿变换,从而将转换至对应位姿的参考对象模型的点云数据带入服务视频提取了客服对象之后的3D点云数据中,以形成参考视频。参考对象模型是与客服对象对应的虚拟现实客服的人物模型,不限于是具备客服对象特征和参考形象特征的人物模型。According to the pose parameters of each key frame, the reference object model is transformed into a pose, so that the point cloud data converted to the reference object model corresponding to the pose is brought into the 3D point cloud data after the customer service object is extracted from the service video. to form a reference video. The reference object model is the character model of the virtual reality customer service corresponding to the customer service object, and is not limited to the character model having the characteristics of the customer service object and the reference image.
对包括参考对象模型的点云数据的参考视频的调整,不限于是参考视频的光流参数与服务视频的光流参数的比对,对参考对象模型的位姿进行调整,以得到光流参数满足预设条件的虚拟现实服务视频。The adjustment of the reference video including the point cloud data of the reference object model is not limited to the comparison of the optical flow parameters of the reference video and the optical flow parameters of the service video. The pose of the reference object model is adjusted to obtain the optical flow parameters. Virtual reality service videos that meet preset conditions.
通过本公开实施例,由于通过服务视频中客服对象的位姿参数,对参考对象模型进行转换以利用转换后的参考对象模型带入服务视频的3D点云数据中形成参考视频,从而对参考视 频的光流参数进行调整以得到虚拟现实服务视频,基于客服对象的服务视频生成对应的虚拟现实服务视频,利用虚拟现实服务视频进行客服服务,打破了同一客服对象在同一时间只能为一个用户提供客服服务的限制,利用虚拟现实服务视频为多个用户同时提供与客服服务。因此,可以解决线上视频服务的服务效率较低的问题,达到提高视频服务的服务效率的技术效果。Through the embodiments of the present disclosure, the reference object model is converted through the pose parameters of the customer service object in the service video, and the converted reference object model is brought into the 3D point cloud data of the service video to form a reference video, so that the reference view is The optical flow parameters of the frequency are adjusted to obtain the virtual reality service video. The corresponding virtual reality service video is generated based on the service video of the customer service object. The virtual reality service video is used for customer service, breaking the problem that the same customer service object can only serve one user at the same time. To limit the provision of customer service services, use virtual reality service videos to provide customer service services to multiple users at the same time. Therefore, the problem of low service efficiency of online video services can be solved, and the technical effect of improving the service efficiency of video services can be achieved.
作为一种可选的实施方式,上述在根据对象位姿参数对参考对象模型进行转换之前,还包括:As an optional implementation, before converting the reference object model according to the object pose parameters, the above further includes:
S11,获取客服对象的面部数据;S11, obtain the facial data of the customer service object;
S12,利用神经网络,将客服对象的面部数据和参考形象特征转化为参考对象模型。S12, use neural network to convert the facial data and reference image features of the customer service object into a reference object model.
参考对象模型不限于通过神经网络,将克服对象的面部数据和参考形象特征生成参考对象模型。客服对象的面部数据不限于为客服对象的面部特征数据,参考形象特征不限于为预设的身体形象特征,不限于包括发型、肢体、服饰、配饰等除面部以外的其余人物形象特征。通过神经网络,利用客服对象的面部数据和预设的参考形象特征生成与客服对象对应的虚拟现实(VR)客服对象,利用VR客服对象进一步生成VR服务视频,以通过VR服务视频替代客服对象进行视频服务的形式,提高视频服务的服务效率。The reference object model is not limited to the neural network, and the reference object model will be generated by overcoming the object's facial data and reference image features. The facial data of the customer service object is not limited to the facial feature data of the customer service object, and the reference image features are not limited to the preset body image features, including hairstyle, limbs, clothing, accessories and other character features other than the face. Through the neural network, the facial data of the customer service object and the preset reference image characteristics are used to generate a virtual reality (VR) customer service object corresponding to the customer service object, and the VR customer service object is used to further generate a VR service video, so that the VR service video can replace the customer service object. The form of video service improves the service efficiency of video services.
作为一种可选的实施方式,上述S204根据服务视频的光流参数,从服务视频包括的各个对象中确定出客服对象,包括:As an optional implementation manner, the above-mentioned S204 determines the customer service object from each object included in the service video according to the optical flow parameters of the service video, including:
S204-1,根据服务视频的光流参数对3D点云数据进行聚类,得到服务视频中包括的各个对象;S204-1, cluster the 3D point cloud data according to the optical flow parameters of the service video to obtain each object included in the service video;
S204-2,利用服务视频中的关键帧图像对各个对象进行识别,以从各个对象中确定出客服对象。S204-2: Use the key frame images in the service video to identify each object to determine the customer service object from each object.
获取服务视频中每个关键帧各自对应的光流参数,通过相邻关键帧之间光流参数的连续性,对服务视频的3D点云数据进行聚类,从而得到服务视频中包括的各个对象。对聚类得到的各个对象进行对象识别,从而确定出客服对象。Obtain the optical flow parameters corresponding to each key frame in the service video, and cluster the 3D point cloud data of the service video through the continuity of the optical flow parameters between adjacent key frames to obtain each object included in the service video. . Object recognition is performed on each object obtained by clustering to determine the customer service object.
作为一种可选的实施方式,上述根据服务视频的光流参数对3D点云数据进行聚类,得到服务视频中包括的各个对象,包括:As an optional implementation, the above-mentioned clustering of 3D point cloud data is performed based on the optical flow parameters of the service video to obtain various objects included in the service video, including:
S204-11,确定服务视频中相邻关键帧的光流参数;S204-11, determine the optical flow parameters of adjacent key frames in the service video;
S204-12,将光流参数与3D点云数据映射形成的二维坐标进行匹配;S204-12, match the optical flow parameters with the two-dimensional coordinates formed by 3D point cloud data mapping;
S204-13,利用相邻关键帧的光流参数的连续性,聚类得到各个对象。S204-13, utilize the continuity of optical flow parameters of adjacent key frames to obtain each object through clustering.
将每一关键帧的光流参数与3D点云数据映射形成的二维坐标进行匹配,利用同一对象在相邻关键帧中光流参数的连续性,聚类得到服务视频中包括的各个3D对象。将3D点云数据映射到二维平面上,从而在映射形成的二维平面上进行光流参数连续性的确定与匹配,不限于设定光流参数连续性的判断条件,从而判断相邻关键帧中,哪些点云数据的光流是连续的,从而根据光流的连续性聚类出各个3D对象。Match the optical flow parameters of each key frame with the two-dimensional coordinates formed by 3D point cloud data mapping, and use the continuity of the optical flow parameters of the same object in adjacent key frames to cluster each 3D object included in the service video . Map 3D point cloud data to a two-dimensional plane to determine and match the continuity of optical flow parameters on the two-dimensional plane formed by mapping. It is not limited to setting the judgment conditions for the continuity of optical flow parameters, so as to judge adjacent key points. In the frame, the optical flow of which point cloud data is continuous, so that each 3D object can be clustered according to the continuity of the optical flow.
作为一种可选的实施方式,上述利用服务视频中的关键帧图像对各个对象进行识别,包括:As an optional implementation, the above method uses key frame images in the service video to identify each object, including:
S204-21,根据聚类结果从服务视频中分割出关键帧图像;S204-21, segment key frame images from the service video according to the clustering results;
S204-22,对关键帧图像进行对象识别,得到识别结果;S204-22, perform object recognition on the key frame image and obtain the recognition result;
S204-23,将识别结果与各个对象进行匹配,并根据匹配结果为各个对象添加对象标签。 S204-23: Match the recognition results with each object, and add object labels to each object according to the matching results.
聚类结果不限于为聚类得到的服务视频中包括的各个3D对象,根据聚类结果从服务视频中分割出关键帧图像,不限于是从服务视频的各个关键帧图像中确定出用于对象识别的关键帧图像,例如与聚类结果在对象数量上、形状上等任意一个或多个维度最匹配的关键帧图像。The clustering result is not limited to each 3D object included in the service video obtained by clustering. The key frame image is segmented from the service video according to the clustering result. It is not limited to determining the object for each key frame image from the service video. The identified keyframe image, for example, the keyframe image that best matches the clustering result in any one or more dimensions such as the number of objects, shape, etc.
对关键帧图像进行物体识别,以得到物体识别结果,将物体识别结果与聚类结果得到的各个对象进行一一匹配,从而根据匹配到的物体识别结果为各个对象添加对象标签,通过对象标签标示聚类得到的各个对象。物体识别结果不限于识别出关键帧图像中每个物体的具体分类,例如人物、话筒、展板等。Perform object recognition on the key frame image to obtain the object recognition result, match the object recognition result with each object obtained from the clustering result one by one, and then add object labels to each object according to the matched object recognition results, and mark them with object labels Each object obtained by clustering. The object recognition results are not limited to identifying the specific classification of each object in the key frame image, such as people, microphones, exhibition boards, etc.
作为一种可选的实施方式,上述S206从3D点云数据中提取出客服对象的点云数据,并确定客服对象的对象位姿参数,包括:As an optional implementation manner, the above-mentioned S206 extracts the point cloud data of the customer service object from the 3D point cloud data, and determines the object pose parameters of the customer service object, including:
S206-1,根据对象标签从3D点云数据中提取出客服对象的点云数据;S206-1, extract the point cloud data of the customer service object from the 3D point cloud data according to the object label;
S206-2,根据客服对象的点云数据,计算得到客服对象的对象位姿以及对象位姿对应的概率。S206-2: Based on the point cloud data of the customer service object, calculate the object pose of the customer service object and the probability corresponding to the object pose.
在为聚类得到的每个3D对象添加对象标签后,不限于通过对象标签确定3D对象中包括的客服对象,例如将标识为“人物”的3D对象确定为客服对象,并从服务视频的3D点云数据中提取客服对象的点云数据。After adding object labels to each 3D object obtained by clustering, the customer service objects included in the 3D objects are not limited to being determined through the object labels. For example, the 3D objects identified as "characters" are determined as customer service objects, and the 3D objects of the service videos are determined. Extract point cloud data of customer service objects from point cloud data.
在提取出客服对象的点云数据的情况下,计算客服对象的对象位姿以及对象位姿对应的概率。客服对象的对象位姿不限于是与服务视频的各个关键帧对应的客服对象的对象位姿,以及对象位姿的概率变化区间。对象位姿不限于是客服对象的3D位姿。When the point cloud data of the customer service object is extracted, the object pose of the customer service object and the probability corresponding to the object pose are calculated. The object pose of the customer service object is not limited to the object pose of the customer service object corresponding to each key frame of the service video, and the probability change interval of the object pose. The object pose is not limited to the 3D pose of the customer service object.
作为一种可选的实施方式,上述S210调整参考视频直至参考视频的光流参数满足预设条件的情况下,得到与客服对象对应的虚拟现实服务视频,包括:As an optional implementation manner, the above-mentioned S210 adjusts the reference video until the optical flow parameters of the reference video meet the preset conditions, and obtains the virtual reality service video corresponding to the customer service object, including:
S210-1,计算参考视频中参考关键帧的光流参数;S210-1, calculate the optical flow parameters of the reference key frames in the reference video;
S210-2,获取服务视频中与参考关键帧对应原始关键帧的光流参数;S210-2, obtain the optical flow parameters of the original key frame corresponding to the reference key frame in the service video;
S210-3,比较参考关键帧的光流参数和原始关键帧的光流参数,得到光流差异参数;S210-3, compare the optical flow parameters of the reference key frame and the optical flow parameters of the original key frame to obtain the optical flow difference parameter;
S210-41,在参考视频的每个参考关键帧对应的光流差异参数均小于预设阈值的情况下,确定得到虚拟现实服务视频;S210-41, when the optical flow difference parameter corresponding to each reference key frame of the reference video is less than the preset threshold, determine that the virtual reality service video is obtained;
S210-42,在光流差异参数大于或等于预设阈值的情况下,根据光流差异参数对参考对象模型的位姿进行调整,直至得到虚拟现实服务视频。S210-42: When the optical flow difference parameter is greater than or equal to the preset threshold, adjust the pose of the reference object model according to the optical flow difference parameter until the virtual reality service video is obtained.
根据客服对象的点云数据计算得到的3D位姿对参考对象模型进行位姿变换,将参考对象模型变换至与3D位姿一致的位姿,并将位姿变换后的参考对象模型的点云数据带入除去客服对象的服务视频的3D点云数据中,形成参考视频。也就是利用位姿变换后的参考对象模型替换服务视频中的客服对象,从而得到包括VR客服的参考视频。Perform pose transformation on the reference object model based on the 3D pose calculated from the point cloud data of the customer service object, transform the reference object model into a pose consistent with the 3D pose, and convert the point cloud of the transformed reference object model into The data is brought into the 3D point cloud data of the service video excluding the customer service object to form a reference video. That is, the reference object model after pose transformation is used to replace the customer service object in the service video, thereby obtaining a reference video including VR customer service.
具体的,参考对象模型的点云数据的带入不限于通过每个关键帧对应的点云数据的带入实现,对每个关键帧实现参考对象模型的带入,以得到各个关键帧中都包括VR客服的参考视频。Specifically, the point cloud data of the reference object model is not limited to the point cloud data corresponding to each key frame. The reference object model is brought in for each key frame to obtain all key frames. Includes reference videos for VR customer service.
在得到参考视频后,获取参考视频中每个参考关键帧的光流参数,并与服务视频的原始关键帧的光流参数进行比较,得到光流差异参数,用于指示参考关键帧与原始关键帧的光流差异。After obtaining the reference video, obtain the optical flow parameters of each reference key frame in the reference video and compare it with the optical flow parameters of the original key frame of the service video to obtain the optical flow difference parameter, which is used to indicate the difference between the reference key frame and the original key frame. Optical flow difference between frames.
在光流差异参数大于或等于预设阈值的情况下,对参考视频进行调整,直至调整后的参考视频的光流差异参数小于预设阈值的情况下,将参考视频作为VR服务视频。对于参考视频 的调整不限于通过调整每个参考关键帧中VR客服的位姿,通过调整带入的VR客服的位姿,使得形成的参考关键帧的光流与原始关键帧的光流的差异降低至预设阈值。When the optical flow difference parameter is greater than or equal to the preset threshold, the reference video is adjusted until the adjusted optical flow difference parameter of the reference video is less than the preset threshold, and the reference video is used as the VR service video. For reference video The adjustment is not limited to adjusting the pose of the VR customer service in each reference key frame. By adjusting the pose of the VR customer service brought in, the difference between the optical flow of the formed reference key frame and the optical flow of the original key frame is reduced to the predetermined level. Set threshold.
具体的,虚拟现实VR服务视频的生成不限于如图3所示。基于包括客服的视频服务,利用SLAM技术提取到视频服务的3D点云数据。同时获取视频服务各个关键帧的光流,利用相邻关键帧的光流连续性进行点云聚类,确定出视频服务中包括的客服的点云数据,进而通过客服的点云数据计算得到每个关键帧中客服的3D位姿。根据3D位姿对客服对应的VR人物模型进行位姿转换后,利用位姿转换后的VR人物模型替换客服带入3D点云数据中,形成参考视。通过参考视频与视频服务的光流比较对参考视频中VR客服进行位姿迭代,直至得到VR服务。Specifically, the generation of virtual reality VR service videos is not limited to that shown in Figure 3. Based on the video service including customer service, SLAM technology is used to extract the 3D point cloud data of the video service. At the same time, the optical flow of each key frame of the video service is obtained, and the optical flow continuity of adjacent key frames is used to perform point cloud clustering to determine the point cloud data of the customer service included in the video service. Then, each customer service point cloud data is calculated. The 3D pose of the customer service agent in key frames. After converting the VR character model corresponding to the customer service according to the 3D pose, the transformed VR character model is used to replace the customer service and bring it into the 3D point cloud data to form a reference view. By comparing the optical flow between the reference video and the video service, the VR customer service in the reference video is iterated until the VR service is obtained.
通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到根据上述实施例的方法可借助软件加必需的通用硬件平台的方式来实现,当然也可以通过硬件,但很多情况下前者是更佳的实施方式。基于这样的理解,本公开的技术方案本质上或者说对现有技术做出贡献的部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质(如只读存储器/随机存取存储器(Read-Only Memory/Random Access Memory,ROM/RAM)、磁碟、光盘)中,包括若干指令用以使得一台终端设备(可以是手机,计算机,服务器,或者网络设备等)执行本公开各个实施例所述的方法。Through the description of the above embodiments, those skilled in the art can clearly understand that the method according to the above embodiments can be implemented by means of software plus the necessary general hardware platform. Of course, it can also be implemented by hardware, but in many cases the former is Better implementation. Based on this understanding, the technical solution of the present disclosure can be embodied in the form of a software product in nature or in part that contributes to the existing technology. The computer software product is stored in a storage medium (such as read-only memory/random access memory). The memory (Read-Only Memory/Random Access Memory, ROM/RAM), magnetic disk, optical disk) includes several instructions to cause a terminal device (which can be a mobile phone, computer, server, or network device, etc.) to execute the disclosure Methods described in various embodiments.
在本实施例中还提供了一种虚拟现实服务视频的生成装置,该装置设置为实现上述实施例及优选实施方式,已经进行过说明的不再赘述。如以下所使用的,术语“模块”可以实现预定功能的软件和/或硬件的组合。尽管以下实施例所描述的装置较佳地以软件来实现,但是硬件,或者软件和硬件的组合的实现也是可能并被构想的。This embodiment also provides a device for generating a virtual reality service video. The device is configured to implement the above embodiments and preferred implementations. What has already been explained will not be described again. As used below, the term "module" may be a combination of software and/or hardware that implements a predetermined function. Although the apparatus described in the following embodiments is preferably implemented in software, implementation in hardware, or a combination of software and hardware, is also possible and contemplated.
图4是根据本公开实施例的虚拟现实服务视频的生成装置的结构框图,如图4所示,该装置包括:Figure 4 is a structural block diagram of a device for generating virtual reality service videos according to an embodiment of the present disclosure. As shown in Figure 4, the device includes:
提取模块41,设置为提取服务视频中的3D点云数据,其中,服务视频为客服对象执行客服服务形成的视频;The extraction module 41 is configured to extract 3D point cloud data in the service video, where the service video is a video formed by the customer service object performing customer service service;
确定模块42,设置为根据服务视频的光流参数,从服务视频包括的各个对象中确定出客服对象;The determination module 42 is configured to determine the customer service object from each object included in the service video according to the optical flow parameters of the service video;
位姿模块43,设置为从3D点云数据中提取出客服对象的点云数据,并确定客服对象的对象位姿参数;The pose module 43 is configured to extract the point cloud data of the customer service object from the 3D point cloud data, and determine the object pose parameters of the customer service object;
转换模块44,设置为根据对象位姿参数对参考对象模型进行转换,并将转换后的参考对象模型的点云数据带入3D点云数据形成参考视频,其中,参考对象模型为客服对象对应的虚拟现实客服;The conversion module 44 is configured to convert the reference object model according to the object pose parameters, and bring the point cloud data of the converted reference object model into the 3D point cloud data to form a reference video, where the reference object model is the corresponding one of the customer service object. virtual reality customer service;
调整模块45,设置为调整参考视频直至参考视频的光流参数满足预设条件,得到与客服对象对应的虚拟现实服务视频。The adjustment module 45 is configured to adjust the reference video until the optical flow parameters of the reference video meet the preset conditions to obtain a virtual reality service video corresponding to the customer service object.
可选的,上述确定模块42包括:根据服务视频的光流参数对3D点云数据进行聚类,得到服务视频中包括的各个对象;利用服务视频中的关键帧图像对各个对象进行识别,以从各个对象中确定出客服对象。Optionally, the above-mentioned determination module 42 includes: clustering 3D point cloud data according to the optical flow parameters of the service video to obtain each object included in the service video; using key frame images in the service video to identify each object to Determine the customer service objects from each object.
可选的,上述确定模块42中根据服务视频的光流参数对3D点云数据进行聚类,得到服务视频中包括的各个对象,还包括:确定服务视频中相邻关键帧的光流参数;将光流参数与 3D点云数据映射形成的二维坐标进行匹配;利用相邻关键帧的光流参数的连续性,聚类得到各个对象。Optionally, the above-mentioned determination module 42 clusters the 3D point cloud data according to the optical flow parameters of the service video to obtain each object included in the service video, and further includes: determining the optical flow parameters of adjacent key frames in the service video; Compare the optical flow parameters with The two-dimensional coordinates formed by 3D point cloud data mapping are matched; the continuity of the optical flow parameters of adjacent key frames is used to cluster each object.
可选的,上述确定模块42中利用服务视频中的关键帧图像对各个对象进行识别,还包括:根据聚类结果从服务视频中分割出关键帧图像;对关键帧图像进行对象识别,得到识别结果;将识别结果与各个对象进行匹配,并根据匹配结果为各个对象添加对象标签。Optionally, the above-mentioned determination module 42 uses the key frame images in the service video to identify each object, which also includes: segmenting the key frame images from the service video according to the clustering results; performing object recognition on the key frame images to obtain the identification Result; match the recognition results with each object, and add object labels to each object based on the matching results.
可选的,上述位姿模块43还包括:根据对象标签从3D点云数据中提取出客服对象的点云数据;根据客服对象的点云数据,计算得到客服对象的对象位姿以及对象位姿对应的概率。Optionally, the above-mentioned pose module 43 also includes: extracting the point cloud data of the customer service object from the 3D point cloud data according to the object label; calculating the object pose and object pose of the customer service object based on the point cloud data of the customer service object. corresponding probability.
可选的,上述虚拟现实服务视频的生成装置还包括模型模块,设置为在根据对象位姿参数对参考对象模型进行转换之前,获取客服对象的面部数据;利用神经网络,将客服对象的面部数据和参考形象特征转化为参考对象模型。Optionally, the above-mentioned generating device of virtual reality service video also includes a model module, which is configured to obtain the facial data of the customer service object before converting the reference object model according to the object pose parameters; and use the neural network to convert the facial data of the customer service object into and reference image features are transformed into reference object models.
可选的,上述调整模块45包括:计算参考视频中参考关键帧的光流参数;获取服务视频中与参考关键帧对应原始关键帧的光流参数;比较参考关键帧的光流参数和原始关键帧的光流参数,得到光流差异参数;在参考视频的每个参考关键帧对应的光流差异参数均小于预设阈值的情况下,确定得到虚拟现实服务视频;在光流差异参数大于或等于预设阈值的情况下,根据光流差异参数对参考对象模型的位姿进行调整,直至得到虚拟现实服务视频。Optionally, the above-mentioned adjustment module 45 includes: calculating the optical flow parameters of the reference key frames in the reference video; obtaining the optical flow parameters of the original key frames corresponding to the reference key frames in the service video; comparing the optical flow parameters of the reference key frames with the original key frames. The optical flow parameter of the frame is obtained to obtain the optical flow difference parameter; when the optical flow difference parameter corresponding to each reference key frame of the reference video is less than the preset threshold, the virtual reality service video is determined to be obtained; when the optical flow difference parameter is greater than or When equal to the preset threshold, the pose of the reference object model is adjusted according to the optical flow difference parameter until the virtual reality service video is obtained.
通过本公开实施例,由于通过服务视频中客服对象的位姿参数,对参考对象模型进行转换以利用转换后的参考对象模型带入服务视频的3D点云数据中形成参考视频,从而对参考视频的光流参数进行调整以得到虚拟现实服务视频,基于客服对象的服务视频生成对应的虚拟现实服务视频,利用虚拟现实服务视频进行客服服务,打破了同一客服对象在同一时间只能为一个用户提供客服服务的限制,利用虚拟现实服务视频为多个用户同时提供与客服服务。因此,可以解决线上视频服务的服务效率较低的问题,达到提高视频服务的服务效率的技术效果。Through the embodiments of the present disclosure, the reference object model is converted through the pose parameters of the customer service object in the service video, and the converted reference object model is brought into the 3D point cloud data of the service video to form a reference video, so that the reference video is The optical flow parameters are adjusted to obtain the virtual reality service video. The corresponding virtual reality service video is generated based on the service video of the customer service object. The virtual reality service video is used for customer service, breaking the problem that the same customer service object can only provide one user at the same time. Limitation of customer service, use virtual reality service video to provide customer service to multiple users at the same time. Therefore, the problem of low service efficiency of online video services can be solved, and the technical effect of improving the service efficiency of video services can be achieved.
需要说明的是,上述各个模块是可以通过软件或硬件来实现的,对于后者,可以通过以下方式实现,但不限于此:上述模块均位于同一处理器中;或者,上述各个模块以任意组合的形式分别位于不同的处理器中。It should be noted that each of the above modules can be implemented through software or hardware. For the latter, it can be implemented in the following ways, but is not limited to this: the above modules are all located in the same processor; or the above modules can be implemented in any combination. The forms are located in different processors.
为便于对本公开所提供的技术方案的理解,下面将结合具体场景的实施例进行详细的阐述。To facilitate understanding of the technical solutions provided by the present disclosure, detailed descriptions will be given below in conjunction with embodiments of specific scenarios.
本公开的实施例还提供了一种计算机可读存储介质,该计算机可读存储介质中存储有计算机程序,其中,该计算机程序被设置为运行时执行上述任一项方法实施例中的步骤。Embodiments of the present disclosure also provide a computer-readable storage medium that stores a computer program, wherein the computer program is configured to execute the steps in any of the above method embodiments when running.
在一个示例性实施例中,上述计算机可读存储介质可以包括但不限于:U盘、只读存储器(Read-Only Memory,简称为ROM)、随机存取存储器(Random Access Memory,简称为RAM)、移动硬盘、磁碟或者光盘等各种可以存储计算机程序的介质。In an exemplary embodiment, the computer-readable storage medium may include but is not limited to: U disk, read-only memory (Read-Only Memory, referred to as ROM), random access memory (Random Access Memory, referred to as RAM) , mobile hard disk, magnetic disk or optical disk and other media that can store computer programs.
本公开的实施例还提供了一种电子装置,包括存储器和处理器,该存储器中存储有计算机程序,该处理器被设置为运行计算机程序以执行上述任一项方法实施例中的步骤。Embodiments of the present disclosure also provide an electronic device, including a memory and a processor. A computer program is stored in the memory, and the processor is configured to run the computer program to perform the steps in any of the above method embodiments.
在一个示例性实施例中,上述电子装置还可以包括传输设备以及输入输出设备,其中,该传输设备和上述处理器连接,该输入输出设备和上述处理器连接。In an exemplary embodiment, the above-mentioned electronic device may further include a transmission device and an input-output device, wherein the transmission device is connected to the above-mentioned processor, and the input-output device is connected to the above-mentioned processor.
本实施例中的具体示例可以参考上述实施例及示例性实施方式中所描述的示例,本实施 例在此不再赘述。For specific examples in this embodiment, reference may be made to the examples described in the above embodiments and exemplary implementations. The example will not be repeated here.
显然,本领域的技术人员应该明白,上述的本公开实施例的各模块或各步骤可以用通用的计算装置来实现,它们可以集中在单个的计算装置上,或者分布在多个计算装置所组成的网络上,它们可以用计算装置可执行的程序代码来实现,从而,可以将它们存储在存储装置中由计算装置来执行,并且在某些情况下,可以以不同于此处的顺序执行所示出或描述的步骤,或者将它们分别制作成各个集成电路模块,或者将它们中的多个模块或步骤制作成单个集成电路模块来实现。这样,本公开不限制于任何特定的硬件和软件结合。Obviously, those skilled in the art should understand that each module or each step of the above-mentioned embodiments of the present disclosure can be implemented by a general computing device, and they can be concentrated on a single computing device, or distributed among multiple computing devices. over a network, they may be implemented with program code executable by a computing device, such that they may be stored in a storage device for execution by the computing device, and in some cases, may be executed in a sequence different from that described here. The steps shown or described may be implemented by fabricating them separately into individual integrated circuit modules, or by fabricating multiple modules or steps among them into a single integrated circuit module. As such, the present disclosure is not limited to any specific combination of hardware and software.
以上所述仅为本公开的优选实施例而已,并不用于限制本公开,对于本领域的技术人员来说,本公开实施例可以有各种更改和变化。凡在本公开实施例的原则之内,所作的任何修改、等同替换、改进等,均应包含在本公开的保护范围之内。 The above are only preferred embodiments of the present disclosure and are not intended to limit the present disclosure. For those skilled in the art, various modifications and changes may be made to the embodiments of the present disclosure. Any modifications, equivalent substitutions, improvements, etc. made within the principles of the embodiments of the present disclosure shall be included in the protection scope of the present disclosure.

Claims (10)

  1. 一种虚拟现实服务视频的生成方法,包括:A method for generating virtual reality service videos, including:
    提取服务视频中的3D点云数据,其中,所述服务视频为客服对象执行客服服务形成的视频;Extract 3D point cloud data in the service video, where the service video is a video formed by the customer service object performing customer service service;
    根据所述服务视频的光流参数,从所述服务视频包括的各个对象中确定出所述客服对象;Determine the customer service object from each object included in the service video according to the optical flow parameters of the service video;
    从所述3D点云数据中提取出所述客服对象的点云数据,并确定所述客服对象的对象位姿参数;Extract the point cloud data of the customer service object from the 3D point cloud data, and determine the object pose parameters of the customer service object;
    根据所述对象位姿参数对参考对象模型进行转换,并将转换后的参考对象模型的点云数据带入所述3D点云数据形成参考视频,其中,所述参考对象模型为所述客服对象对应的虚拟现实客服;The reference object model is converted according to the object pose parameters, and the point cloud data of the converted reference object model is brought into the 3D point cloud data to form a reference video, wherein the reference object model is the customer service object Corresponding virtual reality customer service;
    调整所述参考视频直至所述参考视频的光流参数满足预设条件,得到与所述客服对象对应的虚拟现实服务视频。The reference video is adjusted until the optical flow parameters of the reference video meet the preset conditions, and a virtual reality service video corresponding to the customer service object is obtained.
  2. 根据权利要求1所述的方法,其中,根据所述服务视频的光流参数,从所述服务视频包括的各个对象中确定出所述客服对象,包括:The method according to claim 1, wherein the customer service object is determined from various objects included in the service video according to the optical flow parameters of the service video, including:
    根据所述服务视频的光流参数对所述3D点云数据进行聚类,得到所述服务视频中包括的所述各个对象;Cluster the 3D point cloud data according to the optical flow parameters of the service video to obtain the various objects included in the service video;
    利用所述服务视频中的关键帧图像对所述各个对象进行识别,以从所述各个对象中确定出所述客服对象。The key frame images in the service video are used to identify each object, so as to determine the customer service object from each object.
  3. 根据权利要求2所述的方法,其中,根据所述服务视频的光流参数对所述3D点云数据进行聚类,得到所述服务视频中包括的所述各个对象,包括:The method according to claim 2, wherein the 3D point cloud data is clustered according to the optical flow parameters of the service video to obtain the various objects included in the service video, including:
    确定所述服务视频中相邻关键帧的光流参数;Determine the optical flow parameters of adjacent key frames in the service video;
    将所述光流参数与所述3D点云数据映射形成的二维坐标进行匹配;Match the optical flow parameters with the two-dimensional coordinates formed by mapping the 3D point cloud data;
    利用所述相邻关键帧的所述光流参数的连续性,聚类得到所述各个对象。The respective objects are obtained by clustering using the continuity of the optical flow parameters of the adjacent key frames.
  4. 根据权利要求3所述的方法,其中,利用所述服务视频中的关键帧图像对所述各个对象进行识别,包括:The method according to claim 3, wherein identifying the respective objects using key frame images in the service video includes:
    根据聚类结果从所述服务视频中分割出所述关键帧图像;Segment the key frame image from the service video according to the clustering result;
    对所述关键帧图像进行对象识别,得到识别结果;Perform object recognition on the key frame image to obtain recognition results;
    将所述识别结果与所述各个对象进行匹配,并根据匹配结果为所述各个对象添加对象标签。The recognition results are matched with the respective objects, and object tags are added to the respective objects according to the matching results.
  5. 根据权利要求1所述的方法,其中,从所述3D点云数据中提取出所述客服对象的点云数据,并确定所述客服对象的对象位姿参数,包括:The method according to claim 1, wherein extracting the point cloud data of the customer service object from the 3D point cloud data and determining the object pose parameters of the customer service object includes:
    根据对象标签从所述3D点云数据中提取出所述客服对象的点云数据;Extract the point cloud data of the customer service object from the 3D point cloud data according to the object tag;
    根据所述客服对象的点云数据,计算得到所述客服对象的对象位姿以及所述对象位姿对应的概率。 According to the point cloud data of the customer service object, the object pose of the customer service object and the probability corresponding to the object pose are calculated.
  6. 根据权利要求1所述的方法,其中,在根据所述对象位姿参数对参考对象模型进行转换之前,还包括:The method according to claim 1, wherein before converting the reference object model according to the object pose parameters, it further includes:
    获取所述客服对象的面部数据;Obtain the facial data of the customer service object;
    利用神经网络,将所述客服对象的面部数据和参考形象特征转化为所述参考对象模型。Using a neural network, the facial data and reference image features of the customer service object are converted into the reference object model.
  7. 根据权利要求1所述的方法,其中,调整所述参考视频直至所述参考视频的光流参数满足预设条件的情况下,得到与所述客服对象对应的虚拟现实服务视频,包括:The method according to claim 1, wherein adjusting the reference video until the optical flow parameters of the reference video meet preset conditions to obtain the virtual reality service video corresponding to the customer service object includes:
    计算所述参考视频中参考关键帧的光流参数;Calculate optical flow parameters of reference key frames in the reference video;
    获取所述服务视频中与所述参考关键帧对应原始关键帧的光流参数;Obtain the optical flow parameters of the original key frames corresponding to the reference key frames in the service video;
    比较所述参考关键帧的光流参数和所述原始关键帧的光流参数,得到光流差异参数;Compare the optical flow parameters of the reference key frame and the optical flow parameters of the original key frame to obtain optical flow difference parameters;
    在所述参考视频的每个参考关键帧对应的光流差异参数均小于预设阈值的情况下,确定得到所述虚拟现实服务视频;When the optical flow difference parameter corresponding to each reference key frame of the reference video is less than the preset threshold, it is determined that the virtual reality service video is obtained;
    在所述光流差异参数大于或等于所述预设阈值的情况下,根据所述光流差异参数对所述参考对象模型的位姿进行调整,直至得到所述虚拟现实服务视频。When the optical flow difference parameter is greater than or equal to the preset threshold, the pose of the reference object model is adjusted according to the optical flow difference parameter until the virtual reality service video is obtained.
  8. 一种虚拟现实服务视频的生成装置,包括:A device for generating virtual reality service videos, including:
    提取模块,设置为提取服务视频中的3D点云数据,其中,所述服务视频为客服对象执行客服服务形成的视频;The extraction module is configured to extract 3D point cloud data in the service video, where the service video is a video formed by the customer service object performing customer service service;
    确定模块,设置为根据所述服务视频的光流参数,从所述服务视频包括的各个对象中确定出所述客服对象;A determination module configured to determine the customer service object from each object included in the service video according to the optical flow parameters of the service video;
    位姿模块,设置为从所述3D点云数据中提取出所述客服对象的点云数据,并确定所述客服对象的对象位姿参数;A pose module configured to extract the point cloud data of the customer service object from the 3D point cloud data, and determine the object pose parameters of the customer service object;
    转换模块,设置为根据所述对象位姿参数对参考对象模型进行转换,并将转换后的参考对象模型的点云数据带入所述3D点云数据形成参考视频,其中,所述参考对象模型为所述客服对象对应的虚拟现实客服;A conversion module configured to convert the reference object model according to the object pose parameters, and bring the point cloud data of the converted reference object model into the 3D point cloud data to form a reference video, wherein the reference object model The virtual reality customer service corresponding to the customer service object;
    调整模块,设置为调整所述参考视频直至所述参考视频的光流参数满足预设条件,得到与所述客服对象对应的虚拟现实服务视频。The adjustment module is configured to adjust the reference video until the optical flow parameters of the reference video meet the preset conditions to obtain the virtual reality service video corresponding to the customer service object.
  9. 一种计算机可读存储介质,所述计算机可读存储介质中存储有计算机程序,其中,所述计算机程序被处理器执行时实现所述权利要求1至7任一项中所述的方法的步骤。A computer-readable storage medium having a computer program stored in the computer-readable storage medium, wherein when the computer program is executed by a processor, the steps of the method described in any one of claims 1 to 7 are implemented. .
  10. 一种电子装置,包括存储器、处理器以及存储在所述存储器上并可在所述处理器上运行的计算机程序,其特征在于,所述处理器执行所述计算机程序时实现所述权利要求1至7任一项中所述的方法的步骤。 An electronic device, including a memory, a processor, and a computer program stored on the memory and executable on the processor, characterized in that when the processor executes the computer program, claim 1 is realized to the steps of the method described in any one of 7.
PCT/CN2023/094580 2022-06-13 2023-05-16 Method and device for generating virtual reality service video, and storage medium WO2023241289A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210663511.6 2022-06-13
CN202210663511.6A CN117274974A (en) 2022-06-13 2022-06-13 Virtual reality service video generation method, device and storage medium

Publications (1)

Publication Number Publication Date
WO2023241289A1 true WO2023241289A1 (en) 2023-12-21

Family

ID=89192189

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/094580 WO2023241289A1 (en) 2022-06-13 2023-05-16 Method and device for generating virtual reality service video, and storage medium

Country Status (2)

Country Link
CN (1) CN117274974A (en)
WO (1) WO2023241289A1 (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110674794A (en) * 2018-11-08 2020-01-10 郭娜 Panoramic dance action modeling method and dance teaching auxiliary system
CN113868472A (en) * 2021-10-18 2021-12-31 深圳追一科技有限公司 Method for generating digital human video and related equipment
CN114422862A (en) * 2021-12-24 2022-04-29 上海浦东发展银行股份有限公司 Service video generation method, device, equipment, storage medium and program product
KR20220064711A (en) * 2020-11-12 2022-05-19 주식회사 엘지유플러스 Method and apparatus for real-time output of user object

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110674794A (en) * 2018-11-08 2020-01-10 郭娜 Panoramic dance action modeling method and dance teaching auxiliary system
KR20220064711A (en) * 2020-11-12 2022-05-19 주식회사 엘지유플러스 Method and apparatus for real-time output of user object
CN113868472A (en) * 2021-10-18 2021-12-31 深圳追一科技有限公司 Method for generating digital human video and related equipment
CN114422862A (en) * 2021-12-24 2022-04-29 上海浦东发展银行股份有限公司 Service video generation method, device, equipment, storage medium and program product

Also Published As

Publication number Publication date
CN117274974A (en) 2023-12-22

Similar Documents

Publication Publication Date Title
US10963679B1 (en) Emotion recognition in video
US11238272B2 (en) Method and apparatus for detecting face image
CN113420719B (en) Method and device for generating motion capture data, electronic equipment and storage medium
CN109903392B (en) Augmented reality method and apparatus
CN107911643B (en) Method and device for showing scene special effect in video communication
CN104392220A (en) Three-dimensional face recognition airport security inspection method based on cloud server
CN115049799A (en) Method and device for generating 3D model and virtual image
CN114723888B (en) Three-dimensional hair model generation method, device, equipment, storage medium and product
CN112395979A (en) Image-based health state identification method, device, equipment and storage medium
WO2020177286A1 (en) Method for establishing social relationships, device and server
CN110648289A (en) Image denoising processing method and device
CN108174141B (en) Video communication method and mobile device
CN113591566A (en) Training method and device of image recognition model, electronic equipment and storage medium
CN113658324A (en) Image processing method and related equipment, migration network training method and related equipment
CN113223125B (en) Face driving method, device, equipment and medium for virtual image
CN113408590B (en) Scene recognition method, training method, device, electronic equipment and program product
CN114399424A (en) Model training method and related equipment
CN113642481A (en) Recognition method, training method, device, electronic equipment and storage medium
CN114120413A (en) Model training method, image synthesis method, device, equipment and program product
WO2023241289A1 (en) Method and device for generating virtual reality service video, and storage medium
CN105323143A (en) Network information pushing method, apparatus and system based on instant message
CN114862716A (en) Image enhancement method, device and equipment for face image and storage medium
CN111738087B (en) Method and device for generating face model of game character
CN114500900A (en) Method and device for searching lost object
CN111950352A (en) Hierarchical face clustering method, system, equipment and storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23822859

Country of ref document: EP

Kind code of ref document: A1