WO2023134269A1 - Display device, and virtual fitting system and method - Google Patents

Display device, and virtual fitting system and method Download PDF

Info

Publication number
WO2023134269A1
WO2023134269A1 PCT/CN2022/128392 CN2022128392W WO2023134269A1 WO 2023134269 A1 WO2023134269 A1 WO 2023134269A1 CN 2022128392 W CN2022128392 W CN 2022128392W WO 2023134269 A1 WO2023134269 A1 WO 2023134269A1
Authority
WO
WIPO (PCT)
Prior art keywords
user
image
display device
image data
data associated
Prior art date
Application number
PCT/CN2022/128392
Other languages
French (fr)
Chinese (zh)
Inventor
黄玖法
Original Assignee
海信视像科技股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 海信视像科技股份有限公司 filed Critical 海信视像科技股份有限公司
Publication of WO2023134269A1 publication Critical patent/WO2023134269A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/06Buying, selling or leasing transactions
    • G06Q30/0601Electronic shopping [e-shopping]
    • G06Q30/0641Shopping interfaces
    • G06Q30/0643Graphical representation of items or shoppers
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/06Buying, selling or leasing transactions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/06Buying, selling or leasing transactions
    • G06Q30/0601Electronic shopping [e-shopping]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/006Mixed reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02PCLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
    • Y02P90/00Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
    • Y02P90/30Computing systems specially adapted for manufacturing

Definitions

  • the present application relates to the technical field of intelligent display devices, in particular to a display device, a virtual fitting system and a method.
  • the virtual fitting system is a service platform that integrates hardware and software. It can build a virtual model through augmented reality (Augmented Reality, AR), artificial intelligence (AI), 3D vision and other technologies, and generate a display based on the virtual model. picture.
  • the virtual fitting system can realize the 360-degree natural fit of virtual clothes and the effect of clothes moving with people. It can be widely used in online shopping, daily wear and other life scenes.
  • the virtual fitting system can be built into smart devices such as smart terminals and display devices.
  • the smart device When using the virtual fitting system, the smart device is controlled to run the application of the virtual fitting system. Then input image data such as photos, then the virtual fitting system can form a synthetic effect map through image synthesis, and display it through smart devices to achieve the purpose of virtual fitting.
  • this virtual fitting method which forms effect images through image synthesis, is only suitable for static image display, and cannot display the fitting effect in real time, nor can it display the effects of various wearing postures and viewing angles.
  • the final effect of this kind of virtual fitting system is poor, and the degree of restoration of fitting is low.
  • the present disclosure provides a display device, including: a display, a camera, a communicator, and a controller.
  • the display is configured to display images and/or user interfaces;
  • the camera is configured to collect image data in real time, and the image data includes images associated with the user;
  • the communicator is configured to establish a communication connection with the server, and the server has a built-in model reconstruction application, using The human body model is generated according to the image data associated with the user;
  • the controller is connected with the display, the camera, and the communicator, and the controller is configured to perform: obtaining the image data associated with the user; sending the image data associated with the user to the server for Make the server generate a human body model according to the image data associated with the user and send the human body model to the controller; add clothing materials to the human body model to synthesize the rendering model; extract action parameters from the image data associated with the user, and according to the The action parameters adjust the model pose of the rendered model to render a fitting picture.
  • the present disclosure also provides a virtual fitting system, including: a display device, an image acquisition device, and a server; wherein, the image acquisition device is connected to the display device, and the display device establishes a communication connection with the server; the image acquisition device is configured to collect image data in real time, The image data includes images associated with the user, and image signal processing is performed on the images associated with the user to generate image data associated with the user; the image acquisition device is also configured to send the image data associated with the user to the display device; the display device It is configured to obtain the image data associated with the user, and send the image data associated with the user to the server; a model reconstruction application is built in the server, and the server is configured to receive the image data associated with the user, and run the model reconstruction application; according to the The image data associated with the user generates a human body model, and the server is also configured to send the human body model to the display device; the display device is also configured to add clothing materials to the human body model to synthesize the rendered model; and, extract from the image data associated with the
  • the present disclosure also provides a virtual fitting method, which is applied to a virtual fitting system.
  • the virtual fitting system includes: a display device, an image acquisition device, and a server; wherein, the image acquisition device is connected to the display device, and the display device establishes a communication connection with the server;
  • the virtual fitting method includes: the image acquisition device collects images associated with the user in real time, and performs image signal processing on the image associated with the user to generate image data associated with the user; the display device obtains the image data associated with the user, and The image data associated with the user is sent to the server; the server receives the image data associated with the user, generates a mannequin based on the image data associated with the user, and sends the mannequin to the display device; the display device adds clothing materials to the mannequin for composite rendering Model; the display device extracts action parameters from the image data associated with the user, and adjusts the model pose of the rendering model according to the action parameters to render a fitting picture.
  • An embodiment of the present disclosure also provides a computer-readable non-volatile storage medium, on which computer instructions are stored, and when the computer instructions are executed by a processor, the computer device executes the above method.
  • FIG. 1 is a scene diagram of a virtual fitting system in an embodiment of the present disclosure
  • FIG. 2 is a hardware configuration diagram of a display device in an embodiment of the present disclosure
  • FIG. 3 is a schematic structural diagram of a home smart wardrobe in an embodiment of the present disclosure
  • FIG. 4 is a schematic structural diagram of a display device with a built-in camera in an embodiment of the present disclosure
  • FIG. 5 is a schematic structural diagram of a display device connected to an external image acquisition device in an embodiment of the present disclosure
  • Fig. 6a is a schematic diagram of a virtual fitting interface in an embodiment of the present disclosure.
  • Fig. 6b is a display effect diagram of clothing option classification in an embodiment of the present disclosure.
  • Fig. 6c is a schematic diagram of an interface for identifying clothes in an embodiment of the present disclosure.
  • Fig. 6d is a schematic diagram of an interface for selecting a clothing color in an embodiment of the present disclosure
  • FIG. 6e is a schematic diagram showing the effect of purchase links in an embodiment of the present disclosure.
  • FIG. 6f is a schematic diagram of a purchase interface displayed in an embodiment of the present disclosure.
  • Fig. 6g is a schematic diagram of the effect of displaying product sales positions in the smart fitting mirror in the embodiment of the present disclosure
  • FIG. 7 is a schematic flow diagram of a virtual fitting method in an embodiment of the present disclosure.
  • FIG. 8 is a software configuration diagram of a server in an embodiment of the present disclosure.
  • FIG. 9 is a schematic diagram of a fitting application interface in an embodiment of the present disclosure.
  • FIG. 10 is a schematic diagram of a key frame-based action driving process in an embodiment of the present disclosure.
  • Fig. 11 is a schematic diagram of a process flow for matching associated clothes in an embodiment of the present disclosure.
  • FIG. 12 is a schematic diagram of an expression matching process in an embodiment of the present disclosure.
  • Fig. 13 is a sequence diagram of data interaction in the virtual fitting system in an embodiment of the present disclosure.
  • the virtual fitting system is a service platform integrating hardware and software, which can construct a virtual model through augmented reality (Augmented Reality, AR), artificial intelligence (Artificial Intelligence, AI) and 3D vision technologies. , and generate a display screen based on the virtual model.
  • augmented reality Augmented Reality, AR
  • artificial intelligence Artificial Intelligence, AI
  • 3D vision technologies 3D vision technologies.
  • Fig. 1 is an exemplary use scene diagram of the virtual fitting system in the embodiment of the present application of some embodiments of the present disclosure.
  • the virtual fitting system provided by the present disclosure may include a control device 100 , a display device 200 , an intelligent terminal 300 , and a server 400 .
  • the virtual fitting system can realize the virtual fitting function through the collaborative work among multiple devices.
  • the control device 100 and the smart terminal 300 can be used for user interaction, and can input control instructions to the display device 200 based on the virtual fitting user interface provided by the display device 200 .
  • Both the display device 200 and the server 400 have data processing capability.
  • the display device 200 is deployed locally, the server 400 is deployed in the cloud, and the display device 200 and the server 400 can exchange data.
  • the control device 100 can receive the user's input operation instructions, and convert the operation instructions into instructions that the display device 200 can recognize and respond to, and play an intermediary role between the user and the display device 200 .
  • the control device 100 may be a remote controller, and the communication between the remote controller and the display device 200 includes infrared protocol communication, bluetooth protocol communication, and other short-distance communication methods, and the display device 200 is controlled wirelessly or wiredly.
  • the user can control the display device 200 by inputting user commands through buttons on the remote control, voice input, control panel input, and the like.
  • a smart terminal 300 (such as a mobile terminal, a tablet computer, a computer, a notebook computer, etc.) can also be used to control the display device 200 .
  • the display device 200 is controlled using an application program running on a smart terminal.
  • the display device may not use the above-mentioned smart terminal or control device to receive instructions, but may receive user control through touch or gesture.
  • the display device 200 can also be controlled in a manner other than the control device 100 and the smart terminal 300.
  • the module for acquiring voice commands configured inside the display device 200 can directly receive the user's voice command control , the user's voice command control can also be received through the voice control device installed outside the display device 200.
  • FIG. 2 is an exemplary hardware configuration diagram of a display device in some embodiments of the present disclosure.
  • the display device 200 includes one or more combinations of functional modules such as a power supply 210 , a communicator 220 , a memory 230 , an interface module 240 , a controller 250 , and a display 260 .
  • the power supply 210 can supply power to the display device 200 so that each functional module can be powered on and run.
  • the communicator 220 is used to establish a communication connection relationship between the display device 200 and the server 400, for example, the display device 200 communicates through a local area network (LAN), a wireless local area network (WLAN) and other networks.
  • the server 400 may provide various contents and interactions to the display device 200 .
  • the memory 230 is used to store various information and application data.
  • the interface module 240 is used to connect the display device 200 with peripheral devices to realize the input or output of specific types of signals.
  • the controller 250 controls the operation of the display device 200 and responds to user's operations by running various software control programs stored in the memory 230 .
  • the display 260 is used to present a user interface, so that the display device 200 has a screen display function, and the display device 200 can present a specific display screen on the display 260 by running the application program in the memory 230 and displaying the process. For example, playback interface, user interface, application program interface, etc.
  • the display 260 may take different forms and have different display ratios.
  • the display 260 may have a shape corresponding to a standard display screen ratio.
  • the display resolution of a smart TV is 3840 ⁇ 2160; the display resolution of a personal computer is 1920 ⁇ 1080; the display resolution of a display terminal is 2400 ⁇ 1080.
  • the displays 260 included in them can be designed in a shape and proportion suitable for actual purposes.
  • the display 260 of the virtual reality device includes two square screens on the left and right;
  • FIG. 3 shows a schematic structural diagram of a home smart wardrobe in some embodiments of the present disclosure. As shown in FIG. Elongated display equal to the width and height of the wardrobe door.
  • the display device 200 should be able to acquire a portrait frame associated with the user.
  • the display device 200 can acquire portrait images through the built-in image acquisition module 270 , that is, the display device 200 further includes the image acquisition module 270 on the basis of including the above-mentioned various functional modules.
  • FIG. 4 shows a schematic structural diagram of a display device with a built-in camera in some embodiments of the present disclosure.
  • the image acquisition module is a camera disposed on the top or bottom of the display device 200 .
  • the display device 200 can acquire portrait images through an external image acquisition device 500 .
  • FIG. 5 shows a schematic structural diagram of a display device externally connected to an image acquisition device in some embodiments of the present disclosure, that is, as shown in FIG. 5 , the display device 200 can be connected to the image acquisition device 500 through the interface module 240 .
  • the image acquisition device 500 has a built-in camera and a transmission circuit, and can take pictures of the user through the camera, and then send the captured image or video to the display device 200 for display through the transmission circuit and the interface module 240 .
  • the display device 200 acts as a direct interaction device for the user, which can receive the user's control instructions, and perform data processing according to the control instructions to form a user interface containing different contents, which are presented through the display 260 .
  • the display device 200 can be used as a dedicated device of the virtual fitting system, that is, the display device 200 is only used to run the virtual fitting program and present the virtual fitting interface.
  • the display device 200 can be applied to a robot assistant in a shopping mall environment, and the robot assistant can perform voice interaction with the user to realize a virtual fitting function.
  • the display device 200 can also be an implementation device of the virtual fitting system, that is, the display device 200 has many functions, and the virtual fitting function is one of the many functions.
  • the display device 200 may be a personal computer, and a virtual fitting application may be installed in the personal computer so that it can implement a virtual fitting function.
  • various application programs can be installed in the display device 200 for realizing specific functions.
  • the installed applications can be system applications or third-party applications.
  • a "virtual fitting" application program for users to download and install may be displayed in the application store provided by the operator of the display device 200 .
  • Clothing material can be pre-built in the "virtual fitting" application program, then the display device 200 can run the application program in response to the user's input operation, and receive the image associated with the user, and the display device 200 can compare the image associated with the user with the pre-installed
  • the constructed clothing material is synthesized, and the synthesis effect is displayed to achieve the purpose of fitting.
  • the fitting application program can be an independent application program or a functional module integrated in a specific application.
  • the display device 200 may display a control or icon for activating the "virtual fitting" function in the shopping application.
  • the shopping application can display a prompt interface to the user, and the prompt interface includes controls or icons used to guide the user to input an image associated with the user.
  • the display device 200 invokes the current material model with the purchased clothing, thereby using the "virtual fitting" function to synthesize the image associated with the user with the material model, and output a picture of the fitting effect.
  • the display device 200 may perform image synthesis processing based on the acquired portrait picture associated with the user.
  • the display device 200 synthesizes the portrait image and the clothing image by adding virtual clothing patterns based on the portrait image to realize virtual fitting. For example, after acquiring the portrait picture associated with the user, the display device 200 can perform feature recognition on the portrait picture to identify the wearable position of the portrait pattern, including upper limbs, lower limbs, hands, feet, neck, top of the head, etc. Then extract the virtual clothing material from the clothing material library, so as to add the corresponding clothing material at each dressing position to complete the virtual fitting.
  • the display device 200 may also perform screen synthesis by adding portrait patterns based on virtual clothing patterns.
  • the virtual fitting application can add a portrait display area to the head area corresponding to the virtual clothing pattern, and display the head image in the acquired portrait pattern picture in the display area, thereby synthesizing a virtual fitting picture.
  • the virtual clothing pattern can be obtained through clothing materials stored in the memory 230 or the cloud server 400 .
  • the operator of the virtual fitting system can generate virtual clothing materials by performing multi-angle image shooting, 3D modeling and other processing methods according to the popular clothing currently on sale.
  • the generated virtual clothing material can be stored in the server 400 of the virtual fitting system.
  • the display device 200 starts the virtual fitting application and selects the clothing to be tried on, the display device 200 can request the virtual clothing material from the server 400 according to the user's selection. .
  • the display device 200 may cache the requested virtual clothing material according to the usage of its own memory 230 . Therefore, when using the virtual fitting system subsequently, according to the clothing to be tried on selected by the user, the corresponding virtual clothing material can be first matched in the local cache of the memory 230, and when the local cache includes the corresponding virtual clothing material, the corresponding virtual clothing material can be extracted Virtual clothing material. And when the corresponding virtual clothing material is not matched in the local cache, the corresponding virtual clothing material may be requested to the server 400 again.
  • the virtual clothing material may include parameters with multiple dimensions independent of each other, including: clothing style, color, material, and the like.
  • the display device 200 can present a variety of different clothing styles through the arrangement and combination of parameters in different dimensions.
  • the same virtual clothing material can be used for the same type of clothing, and the appearance of different clothing can be obtained by adjusting parameters such as color and material during virtual try-on. Therefore, the display device 200 can combine multiple clothing materials with less model data, reducing the construction amount of clothing models and the amount of data transmission when requesting clothing materials.
  • the display device 200 may present a virtual fitting interface for the user after running the virtual fitting application.
  • Fig. 6a shows a schematic diagram of a virtual fitting interface of some embodiments of the present disclosure.
  • the virtual fitting interface may include display windows, clothing options, control options, and outfit recommendation options.
  • the virtual fitting interface can display different options to the user, and the display device 200 responds to the user's selection operation on the virtual fitting interface to control and synthesize a virtual fitting image.
  • the portrait image of the user collected by the image acquisition device 500 can be displayed in the display window in real time, and if the selected clothing option is "clothing A", the display device 200 can call the “clothing A” from the local cache or the server 400. ” and display the corresponding virtual clothing material through the display window. Similarly, if the option of "clothing B" is selected among the multiple clothes, the display device 200 may call the virtual clothes material corresponding to "clothes B" and display it through the display window.
  • FIG. 6b shows an exemplary clothing option classification display effect diagram of some embodiments of the present disclosure.
  • the display device 200 can simultaneously display images corresponding to the selected top clothing “top A” and bottom clothing “pants B” in the display window.
  • “Top A” and “Top B” which are both tops, they cannot be selected at the same time, but the clothing material selected later replaces the clothing material selected first.
  • virtual clothing materials of the same category can be further divided into more detailed categories.
  • the classification of tops can be further divided into categories such as spring clothes, summer clothes, autumn clothes, and winter clothes according to the wearing season; it can also be further divided into categories such as coats and shirts according to the wearing position.
  • the display device 200 also needs to call different virtual clothing materials.
  • the display device 200 determines the clothing to be tried on selected by the user by acquiring information such as the product name and item number of the clothing input by the user. For this type of interaction, the display device 200 can search the clothing material library according to the user information, so as to call the clothing material related to the input information.
  • the display device 200 can also identify image information such as display pictures, physical images, and barcodes on hang tags based on image recognition technology to determine the clothing to be tried on.
  • Figure 6c shows a schematic diagram of an interface for identifying clothing in some embodiments of the present disclosure, and the virtual fitting interface may include the option "Identify Clothes", if the option "Identify Clothes" is selected by the user option, the display device 200 can automatically start the image acquisition device 500 to acquire the clothing image, and recognize the clothing image, and call the virtual clothing material according to the image recognition result.
  • the basis for identifying the clothing image can be calculated by calculating the similarity between the captured clothing image and the preset standard image, and the standard image with the highest similarity is used as the recognition result, and stored in the database. Call the virtual clothing material corresponding to the standard image.
  • a clothing option in the virtual fitting interface corresponds to a type of virtual fitting material.
  • the clothing parameters can be adjusted through the control options.
  • FIG. 6d shows a schematic diagram of an interface for selecting the color of clothing in some embodiments of the present disclosure.
  • the display device 200 can display color control options in the virtual fitting interface, so that Show the user the function of selecting any color from the preset multiple color control options, and control the current color of the virtual clothing screen displayed in the display window.
  • control options can also be used to control and adjust the display screen of the display window, for example, rotating the display angle, partially zooming in, adjusting brightness, and beautifying functions.
  • the interactive action is adjusted based on the input parameters matched with the corresponding option selected by the user, and the presentation effect of the display window on the display device 200 is controlled.
  • the outfit recommendation option is used to enable or disable the outfit recommendation function of the virtual fitting application.
  • the clothing recommendation function can automatically display the recommended clothing screen in the display window according to the clothing selected by the user after determining any clothing options selected by the user.
  • the clothing recommendation function can match the clothing materials of unselected categories in the virtual clothing material library according to a specific style algorithm and based on the clothing options selected by the user.
  • the outfit recommendation algorithm can perform unified matching operations based on dimensions such as color, type, and style. For example, in response to the user's operation of selecting black and formal tops, the outfit recommendation algorithm can automatically match black and formal bottoms and shoes, and call the corresponding virtual clothing model, which will be displayed together with the selected top virtual clothing model in the display window.
  • the virtual fitting interface presented by the display device 200 may also include prompt content and link options.
  • the virtual fitting interface can also include a "save outfit" option.
  • the user can click the "save outfit” option to save all items in the current display window. Showcase patterns.
  • the virtual fitting application can also automatically display the purchase link or shopping guide information of the corresponding clothing for the user after responding to the user's click "save outfit", so that the user can purchase the corresponding clothing according to the virtual fitting results.
  • FIG. 6e shows a schematic diagram of displaying purchase link effects in some embodiments of the present disclosure
  • FIG. 6f shows a schematic diagram of displaying purchase interface in some embodiments of the present disclosure.
  • the display device 200 may present a product link corresponding to the virtual clothing pattern displayed in the current display window for the user to select. After obtaining any product link selected by the user, the display device 200 can jump from the virtual fitting interface to the product detail interface.
  • the application prompts the user to display an icon or control for inputting an image associated with the user, responds to the user's operation, opens the file manager, and selects Select a picture file in the save path of the file, and use the picture file as the image input associated with the user.
  • the built-in camera of the display device 200 or the external image acquisition device can also take photos for the user, and use the images obtained by the photos as images associated with the user to input into the application.
  • the virtual fitting application can also provide a dynamic fitting image screen, so that the dynamic fitting image screen can obtain the information associated with the user input by the user by uploading a video file or recording a video file in real time. image.
  • the input video image can be displayed in a specific area in the application interface to obtain a better composite effect.
  • the display device 200 can also use augmented reality technology to add a virtual clothing material model to the video screen uploaded by the user and associated with the user, and to The target is tracked, so that the clothing material model can follow the movement changes of the character target in the video.
  • this dynamic fitting method is limited by the limitations of AR technology, resulting in poor fusion of clothing material and character targets in the video, and clothing cannot accurately fit the character targets.
  • there is a large delay in the display effect of the clothes following the movements of the characters so the interactive experience of the virtual fitting function is poor.
  • the virtual fitting system includes a display device 200 , a server 400 and an image acquisition device 500 .
  • the image collection device 500 is used to collect images associated with the user in real time, and process image signals to form image data associated with the user.
  • a model reconstruction application can be built in the server 400, and the model reconstruction application can generate a human body model according to the image data associated with the user.
  • the server 400 sends the human body model to the display device 200 .
  • the display device 200 is used to run the virtual fitting application program, and render the human body model to synthesize and display the virtual fitting picture.
  • the virtual fitting method includes the following:
  • the image associated with the user is collected in real time by the image collection device 500 .
  • the image acquisition device 500 may include a camera and a data processing module, wherein the camera may capture images of the user's environment to obtain images associated with the user.
  • the data processing module may perform image signal processing on the image associated with the user to generate image data associated with the user, and input the image data associated with the user into the display device 200 as an image signal.
  • the image acquisition device 500 may be a functional module built in the display device 200, and the display device 200 is an integral device.
  • the image acquisition device 500 is a camera on the display device 200, which can be under the unified control of the controller 250 in the display device 200, and can directly send the collected images associated with the user to the controller 250.
  • the camera should be set at a specific position on the display device 200 in order to collect images associated with the user.
  • the camera can be set on the top of the smart TV, and the shooting direction of the camera is the same as the light emitting direction of the screen of the smart TV, so that images associated with the user located in front of the screen can be captured.
  • the image acquisition device 500 can also be an external device connected to the display device 200, that is, as shown in FIG. High Definition Multimedia Interface, HDMI interface), analog or data high-definition component input interface, composite video broadcast signal (Composite Video Broadcast Signal, CVBS) interface, (Universal Serial Bus (Universal Serial Bus, USB) interface and one of the RGB ports can support a specific data transmission method. Then the image acquisition device 500 can send the collected image associated with the user to the display device 200 through the specific data transmission method supported by the interface after accessing the interface module 240. For example, after being connected to the smart TV, the image acquisition device 500 may communicate with the smart TV through Open Natural Interaction (OpenNI), so as to send the collected image associated with the user to the smart TV.
  • OpenNI Open Natural Interaction
  • the image data associated with the user sent by the image acquisition device 500 to the display device 200 may also include image recognition results, bone parameters, expression parameters, and gesture recognition results. data etc. Therefore, in some embodiments, the data processing module of the image acquisition device 500 can also have a built-in image processing application. After the camera captures an image associated with the user, the data processing module can run the image processing application to process the image associated with the user. The image is recognized.
  • the image acquisition device 500 can identify "image-depth" data, 3D human skeleton key point coordinates, portrait head recognition position, portrait target tracking point, etc. in the image associated with the user by running image processing applications with different functions. content. Therefore, when the recognized content is sent to the display device 200 , the display device 200 can be provided with functional supports such as RGBD image data, limb driving, human body tracking, face reconstruction material and expression driving material. Accordingly, these data may, together with user-associated images, constitute user-associated image data.
  • the image acquisition device 500 needs to have specific hardware support.
  • the image acquisition device 500 can be a camera group with multiple lenses, and the multiple lenses can detect the same target from different positions, so that by comparing the image capture results and angle differences of multiple lenses , to calculate the depth information of the target.
  • the image acquisition device 500 can also integrate sensor elements with specific functions, so that the images collected by the device 500 associated with the user include data for three-dimensional modeling.
  • a laser radar component can be built in, and the laser radar can emit laser light to the shooting area, and detect the reflected laser signal to measure the distance between the target in the shooting area and the image acquisition device 500, and compare the scanning result with the image acquisition device 500.
  • Each frame of image captured by the camera is correlated to generate user-associated image data with image depth.
  • FIG. 7 shows a schematic flowchart of a virtual fitting method in some embodiments of the present disclosure.
  • the image acquisition device 500 acquires the image associated with the user and generates the image data associated with the user, it can send the image data associated with the user to the display device 200 .
  • the display device 200 acquires the image data associated with the user as data input for the fitting application.
  • the display device 200 may send an activation command to the image acquisition device 500 after the fitting application is started, control the image acquisition device 500 to start capturing images associated with the user, and perform processing on the images associated with the user to obtain images associated with the user data.
  • the display device 200 may acquire image data associated with the user including different contents from the image acquisition device 500 .
  • the display device 200 may set the type of image data associated with the user to include only image content in the start instruction sent to the image capture device 500 .
  • the image data associated with the user needs to be specified in the start command, including "image-depth" data, 3D human skeleton key point coordinates and other data content.
  • the image forms may also be different.
  • the image data associated with the user in the form of a picture can be obtained from the image acquisition device 500; image data associated with the user.
  • the display device 200 After the display device 200 acquires the image data associated with the user, it can render the fitting result screen according to the image data associated with the user. In order to present a better clothing fusion effect, the display device 200 may obtain a fitting screen based on a human body model. Therefore, after acquiring the image data associated with the user, the display device 200 can create a human body model according to the image data associated with the user. However, due to the huge amount of data processing during the creation of the human body model and the limitation of the hardware configuration of the display device 200, the quality of the human body model locally created by the display device 200 is rough and the accuracy is poor. Therefore, in some embodiments, after acquiring the image data associated with the user, the display device 200 may send the image data associated with the user to the server 400 for modeling processing.
  • a model reconstruction application may be built in the server 400 .
  • a model reconstruction application can generate a mannequin based on image data associated with a user.
  • multiple initial human body models may be pre-configured, and multiple initial human body models may be set according to factors such as age and gender.
  • the server 400 can first identify the image in the image data associated with the user to determine the age, gender and other information of the person in the image, and then based on these information from Call the appropriate one among the preset multiple initial human models.
  • the server 400 can also read data such as "image-depth” data and 3D human skeleton key point coordinates in the image data associated with the user, and set the initial human body model according to the read data, Modify, adjust, so that the initial model gradually has portrait features in the image data associated with the user.
  • the server 400 may extract the proportional relationship between the user's head width and shoulder width from the image data associated with the user, and adjust the head width and shoulder width in the initial model according to the proportional relationship.
  • FIG. 8 shows a software configuration diagram of an exemplary server in some embodiments of the present disclosure.
  • the human body model created by the server 400 may include multiple parts, each part can be parameterized and adjusted independently, and supports replacement through a preset material library, so as to quickly generate the human body model.
  • a variety of hairstyle materials can be pre-stored in the server 400.
  • the server 400 After the server 400 receives the image data associated with the user, it can match similar models in the hairstyle material library according to the portrait hairstyle and color in the image data associated with the user. , and add the hair model to the mannequin's head as the initial hairstyle.
  • the fitting application has the function of allowing the user to arbitrarily choose to change the hairstyle of the character in the virtual fitting effect screen.
  • the display device 200 can then request the server 400 to rebuild. model, the server 400 can select a new hairstyle material from the hairstyle material library and add it to the human body model to reconstruct the human body model.
  • multiple modeling modules can be built in the server 400, including a head reconstruction module, a body reconstruction module, an expression recognition module, a trial hair module, and the like.
  • Each modeling module can construct a corresponding virtual model through a specific model reconstruction method.
  • head modeling can further include geometric reconstruction and texture reconstruction units, wherein the geometric reconstruction unit can perform point cloud denoising processing, triangulation processing, point cloud smoothing processing, and point cloud fusion processing according to the image data associated with the user , so that the shape of the head model gradually tends to be consistent with the person target in the image data associated with the user.
  • the texture reconstruction unit can perform portrait segmentation processing, skin color migration processing, and skin color fusion processing on the head model, so that the appearance of the head model is gradually consistent with the person target in the image data associated with the user.
  • models of body parts can also be generated based on specific modeling methods.
  • a body model can be created through a skinned multi-person linear model (Skinned Multi-Person Linear Model, SMPL). Therefore, the body reconstruction module may include an SMPL model generation unit, a parameterized mapping unit, and a stitching unit.
  • the server 400 can generate an initial model through the SMPL model generation unit, and then extract human body parameters from the image data associated with the user, so as to perform parameter mapping from the SMPL model to the parameterized model, and obtain an image corresponding to the user.
  • the parametric body model of the data, and finally the body model and the head model are spliced through the splicing unit to obtain the human body model.
  • the display device 200 in order to obtain the human body model, when the display device 200 sends the image data associated with the user to the server 400, it can identify the portrait target from the image data associated with the user, and add bones to the portrait target according to the recognition result key points to generate bone parameters. Then send the skeleton parameters to the server 400, so that the server 400 can set the joint point positions of the human body model according to the skeleton parameters.
  • the human body model after adding the bone parameters can change according to the rules corresponding to the bone parameters, so as to simulate a model posture that is more in line with the real state of the character.
  • the built-in processing unit can also be used to establish the expression base model and estimate the expression parameters, so as to set specific expression forms for the human body model according to the expressions of the characters in the image data associated with the user.
  • the preset functional units can be used to set the hair matching function and hair penetration processing function, etc. to create a more realistic hair model.
  • the server 400 Since the server 400 generates or reconstructs the human body model for the subsequent virtual fitting process, so in some embodiments, in addition to the built-in application for generating the human body model, the server 400 may also have a built-in application for generating the clothing model .
  • the server 400 can perform clothing modeling, clothing material simulation, and clothing deformation processing by running the application, so as to create a clothing model based on clothing image data.
  • the display device 200 can also send the clothing image data to the server 400.
  • the clothing image data can be obtained by taking images of the clothing from multiple angles. Get six views of clothing in the product interface of , namely front view, back view, left view, right view, bottom view and top view.
  • the server 400 can automatically model and generate clothes patterns according to these images, and then through rendering processes such as cloth simulation and clothes deformation, a clothes model that conforms to the real effect can be obtained.
  • the server 400 may send the generated human body model to the display device 200 .
  • the display device 200 After receiving the human body model, the display device 200 renders the human body model to render an image with a fitting effect. During the rendering of the human body model, the display device 200 may add clothing materials to the human body model to synthesize a rendered model.
  • the display device 200 may also optimize the human body model sent by the server 400 .
  • the display device 200 can perform head synthesis and restoration processing on the human body model based on the Unity engine, including performing head shape processing through the meshfilter tool and meshrender tool; performing texture processing through the texture tool and normalmap tool; and performing head shape processing through the blendshape/morph tool.
  • the face parameter model algorithm is used to process expression parameter information and adjust the hair to fit the head shape.
  • the display device 200 may also display the image part in the image data associated with the user while sending the image data associated with the user to the server 400 .
  • FIG. 9 shows a schematic diagram of a fitting application interface of some embodiments of the present disclosure.
  • the program interface of a fitting application program can include a fitting window and an original window.
  • the rendering results of the human body model and clothing materials can be displayed in the fitting window, and the image data associated with the user can be displayed in the original window.
  • the image part is the image frame captured by the image acquisition device 500 in real time.
  • the display device 200 After the display device 200 synthesizes the rendered model, it can also extract action parameters from the image data associated with the user, and adjust the model pose of the rendered model according to the action parameters to render a fitting picture. That is, the display device 200 may determine the user's action according to the image data associated with the user, and control the rendering model to follow the user's action corresponding to the image data associated with the user.
  • the actions that can be presented in the fitting screen include body actions, head actions, facial expressions, gesture actions, and clothing actions.
  • the motion parameters can be calculated and obtained by the image acquisition device 500 through head detection, 3D human skeleton key point detection, facial expression detection, and gesture recognition detection.
  • the display device 200 can detect the user's action by comparing multiple frames of images in the image data associated with the user. That is, in some embodiments, in the step of adjusting the model pose of the rendered model according to the action parameters, the display device 200 can traverse the skeleton key points of each frame image in the image data associated with the user; Skeleton key point position to obtain the movement distance of each bone key point to move the joint point position of the human body model according to the movement distance. Through the coordinated adjustment of multiple skeleton key points, the display device 200 can adjust the pose of the person in the rendered model to follow the image content of each frame of the image data associated with the user to the same action, realizing the action following effect.
  • the display device 200 can make the human body model produce a reasonable running mode by skinning, that is, adding bones to the model. For example, after obtaining the human body model, the display device 200 can skin the human body model and the clothing model and provide a texture for the cloth simulation algorithm for skinning, and use the blendshape/morph tool to process the skeleton skinning animation for the model.
  • the virtual fitting method can collect image data associated with the user through the image acquisition device 500 in real time, and generate a human body model through the server 400 according to the image data associated with the user, and display a human body model on the display device 200. Add clothing materials to the model to form a rendered model.
  • the display device 200 can also acquire user actions according to the image data associated with the user, and adjust the model posture of the rendering model in real time according to the user actions, so as to realize avatar driving.
  • the method can realize the effect of 3D, dynamic, real-time and multi-angle display of the fitting screen through the human body model.
  • the server 400 executes model building and reconstruction, which can share the data processing load of the display device 200 and is conducive to generating a more detailed and realistic character model.
  • FIG. 10 shows a schematic diagram of a key frame-based action driving process in some embodiments of the present disclosure.
  • the server 400 sets the human body model parameters according to the image depth parameters, and sends the human body model to the display device 200 .
  • the display device 200 may execute S1004 in subsequent applications.
  • the key frame may be a frame of image corresponding to a specific time point.
  • the display device 200 may extract a frame of image from the image data associated with the user every 1 second as a key frame.
  • the key may also be a frame of image obtained at intervals of a certain number of frames.
  • the display device 200 may extract a frame of image from image data associated with the user at an interval of 20 frames as a key frame.
  • the key frame may also be a frame of image with a distinctive portrait target obtained by performing image recognition on the image frame. For example, by inputting the image data associated with the user into the portrait recognition model frame by frame, when it is determined by the model recognition that the image frame contains a portrait, it will be marked as a key frame.
  • the initial key frame is an image frame used to extract image depth parameters to generate a human body model.
  • the real-time key frame is the key frame extracted by the display device 200 in real time from the image data associated with the user.
  • the image similarity can be determined by calculating the histograms of the two images separately, and then calculating the normalized correlation coefficient of the two histograms, such as Bhattacharyachian distance, histogram intersection distance and other data, to determine the similarity of the two images.
  • the content of the two key frame images is the same, but when the user is in an action state, the content of the two key frame images is different, and the greater the range of motion, the lower the similarity of the content of the two frame images , so the user's action state can be detected by image similarity.
  • the display device 200 may compare the image similarity with a preset similarity threshold.
  • the display device 200 can execute the step of extracting motion parameters from the image data associated with the user, so as to drive the human body model to generate motions according to the motion parameters.
  • the display device 200 can extract the image depth parameter from the real-time key frame image, and The extracted image depth parameters are sent to the server 400, so that the server 400 reconstructs the human body model according to the image depth parameters.
  • the screen content included in the image may gradually vary with the user's actions, that is, the image similarity between the initial key frame and the subsequently extracted key frame gradually decreases. Therefore, In order to maintain the consistency of the display device 200 driving the model action, reduce the number of times of remodeling.
  • the display device 200 may record the initial key frame image after extracting the image depth parameter from the initial key frame image. And after the step of extracting motion parameters from the image data associated with the user is performed, or after the step of sending the image depth parameters to the server, the real-time key frame images are used to replace the recorded initial key frame images.
  • the display device 200 may first extract an initial key frame T0 from image data associated with the user, and send the initial key frame T0 to the server 400 to generate a human body model. Moreover, when the display device 200 subsequently renders the fitting picture of the mannequin, it can continuously acquire the real-time key frame, that is, T1. Then the display device 200 can compare the initial key frame T0 and the real-time key frame T1 after acquiring the real-time key frame T1, and calculate the similarity S01 of the two frames of images.
  • start Action parameters are extracted from the key frame T1 or the image data associated with the user to drive the human body model to generate actions and follow the changes of the user's actions.
  • the display device 200 may use the real-time key frame T1 to replace the initial key frame T0 as the initial key frame in the subsequent motion judgment process. That is, when the real-time key frame T2 is obtained, the similarity between the key frame T1 and the key frame T2 can be compared to continue tracking user actions or reconstructing the human body model.
  • the above embodiment can continuously analyze the image frame associated with the user collected by the image collection device 500 through the initial key frame and the real-time key frame, so as to track the actual action of the user. Moreover, the display device 200 can determine whether it is necessary to reconstruct the human body model by comparing the image similarity between two adjacent key frames, so that the human body model can reduce the number of reconstruction models on the premise that it can be synchronized in time, and improve the fitting class. The responsiveness of the application.
  • the fitting application provides real-time dynamic fitting functions, it can also provide users with outfit recommendations, as shown in Figure 11, which shows the process of matching associated clothes in some embodiments of the present disclosure schematic diagram.
  • the display device 200 may acquire a selection instruction input by the user for selecting clothing.
  • a selection instruction input by the user for selecting clothing.
  • multiple clothing options can be set, and any option in the interface that is selected can be determined by obtaining user interaction operations, so as to input a selection instruction.
  • the user selects multiple target clothes among multiple clothes options, that is, at least one target clothes is specified in the selection command.
  • the selection instruction may be that the user can simultaneously select upload and download to input the selection instruction.
  • the display device 200 may respond to the selection instruction and extract the target clothing material from the clothing material library. Because clothing can be divided into multiple categories, such as tops, bottoms, shoes, hats, bags, etc. These types can be matched with each other to form the final dressing effect. Therefore, when the selection instruction indicates that not all types are selected, the fitting application can automatically match other suitable types of clothing according to the characteristics of the selected clothing, so as to present a better fitting effect.
  • the display device 200 can match the associated clothing material in the clothing material library according to the target clothing material according to the preset dressing recommendation rules.
  • the preset recommendation rules for outfits can be comprehensively set based on categories such as color, purpose, style, and applicable age. For example, when the user chooses blue and white tops with a small and fresh style, related clothing materials such as blue and white bottoms and shoes can be recommended to the user according to the preset recommendation rules for outfitting.
  • the display device 200 may add the target clothing material and the associated clothing material to the human body model, so as to form a rendering model with the effect of dressing in the opinion.
  • the display device 200 can pre-set multiple dressing recommendation rules, and each rule can match the appropriate associated clothing under the rule according to the target clothing material selected by the user. footage and preview it through multiple windows for enhanced display.
  • the display device 200 can quickly implement expression switching in the form of a preset standard expression template. That is, as shown in FIG. 12 , FIG. 12 shows a schematic diagram of an expression matching process in some embodiments of the present disclosure.
  • the display device 200 can identify the head area in the image data associated with the user, and detect the user's expression in the head area, and then according to the expression type of the user's expression, Match the expression model of the same type as the expression in the preset expression library, thereby replacing the facial area in the rendered model with the expression model.
  • the display device 200 may identify the head area in the image associated with the user by detecting the target shape in the image data associated with the user and the screen layout characteristics in the shape. From the head area, use the expression recognition model to recognize the current user's expression.
  • the facial expression recognition model can be obtained by training an artificial intelligence model with sample data. That is, by inputting a large amount of sample data with expression labels into the initial model, and setting the model output result as the classification probability of the image belonging to a specific expression category, and then calculating the error between the classification probability and the expression label, and adjusting according to the error backpropagation Model parameters, so that the output result of the model is gradually the same as the label result, and an expression recognition model is obtained.
  • the expression recognition model After the expression recognition model outputs the classification probability of the current image for a certain expression, the expression with the highest classification probability can be used as the user's expression, such as a smile. Then match the same type of expression model in the preset database according to the recognized user's expression, that is, the standard smile model, and use the standard smile model to replace the facial area in the rendering model, so that the rendering model can display smiling expressions.
  • the above embodiment can quickly replace the facial area in the rendering model by matching the standard expression model, so that the display device 200 does not need to modify the facial parameters of the model to reduce the amount of data processing.
  • better facial expression tracking timeliness can be obtained, and the display effect can be improved.
  • some embodiments of the present disclosure further provide a display device 200 including: a display 260 , a camera, a communicator 220 and a controller 250 .
  • the display 260 is configured to display images and/or user interfaces;
  • the camera is configured to collect image data in real time, and the image data includes images associated with the user;
  • the communicator 220 is configured to establish a communication connection with the server 400, and the server 400 has a built-in model a reconstruction application for generating a mannequin from image data associated with a user;
  • the controller 250 is configured to perform the following procedural steps:
  • FIG. 13 shows an exemplary virtual fitting system data interaction sequence diagram of some embodiments of the present disclosure.
  • the present disclosure also provides a virtual fitting system, including: a display device 200, an image acquisition device 500, and a server 400; wherein, the image acquisition device 500 is connected to the display device 200, and the display device 200 establishes communication with the server 400 connect.
  • the image acquisition device 500 receives an image associated with the user from the user.
  • the image acquisition apparatus 500 performs image signal processing on an image associated with a user, so as to generate image data associated with the user.
  • the image acquisition apparatus 500 sends the image data associated with the user to the display device 200.
  • the display device 200 sends the image data associated with the user to the server 400.
  • the server 400 runs a model reconstruction application; generates a human body model according to the image data associated with the user.
  • the server 400 sends the human body model to the display device 200 .
  • the display device 200 adds clothing materials to the human body model to synthesize a rendered model; and extracts action parameters from image data associated with the user, and adjusts the model pose of the rendered model according to the action parameters to render a fitting picture.
  • the display device 200 sends a fitting screen to the device associated with the user.
  • the entire virtual fitting system may include: a cloud server 400 , a local display device 200 and an image acquisition device 500 .
  • the cloud server 400 can be responsible for human body reconstruction, including basic algorithm modules such as head reconstruction, body reconstruction, expression recognition, hair try-on, etc., and provides functional support for the display device 200 such as human body model, expression driving parameters, and hair try-on.
  • the server 400 may rely on the data collected by the image collection device 500 as input, and output the modeling processing results to the display device 200 to support applications.
  • the image acquisition device 500 is responsible for providing acquisition data, including image signal processing (Image Signal Processing, ISP) debugging, RGBD stream, key points of 3D human skeleton, human head detection, multi-target tracking data, etc., so as to communicate and transmit through OpenNI, Provide the display device 200 with functional support for RGBD image data, limb driving, human body tracking, face reconstruction material, and expression driving material.
  • the display device 200 can be responsible for functions such as rendering of the human body model, display and rendering of clothing materials, motion driving, and local parameter adjustment.
  • the virtual fitting system can collect image data associated with the user in real time during use, and send the image data associated with the user to the server 400 to generate a human body model. Then add clothing materials to the human body model to synthesize the rendering model, and extract the character's actions in real time, so as to adjust the model pose of the rendering model according to the action parameters to form a fitting screen.
  • the virtual fitting system can realize the dynamic 3D virtual fitting function, and display the user's actions in real time through the rendering model, so as to achieve the effect of clothes moving with people and recommended outfits, and solve the problem that the traditional virtual fitting method cannot display the fitting effect in real time.

Abstract

A display device, and a virtual fitting system and method. According to the method, image data associated with a user can be collected in real time during use, and the image data associated with the user can be sent to a server to generate a human model. Clothing materials are added into the human model to synthesize a rendering model, and character motions are extracted in real time, so that a model posture of the rendering model is adjusted according to motion parameters to form a fitting picture. According to the method, a dynamic 3D virtual fitting function can be achieved, and the user motions are displayed in real time by means of the rendering model, thereby achieving the use effects of clothes moving with people and wearing recommendation, and solving the problem in traditional virtual fitting methods that the fitting effect cannot be displayed in real time.

Description

一种显示设备、虚拟试衣系统及方法A display device, virtual fitting system and method
相关申请的交叉引用Cross References to Related Applications
本申请要求在2022年01月17日提交中国专利局、申请号为202210051018.9、申请名称为“一种显示设备、虚拟试衣系统及方法”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of the Chinese patent application with the application number 202210051018.9 and the application title "a display device, virtual fitting system and method" submitted to the China Patent Office on January 17, 2022, the entire contents of which are incorporated by reference in this application.
技术领域technical field
本申请涉及智能显示设备技术领域,尤其涉及一种显示设备、虚拟试衣系统及方法。The present application relates to the technical field of intelligent display devices, in particular to a display device, a virtual fitting system and a method.
背景技术Background technique
虚拟试衣系统是一种集成硬件和软件的服务平台,可以通过增强现实(Augmented Reality,AR)、人工智能(Artificial Intelligence,AI)以及3D视觉等技术,构建虚拟模型,并依据虚拟模型生成显示画面。虚拟试衣系统可以实现虚拟衣物360度自然贴身、衣随人动的效果,可广泛应用于网络购物、日常穿搭等生活场景中。The virtual fitting system is a service platform that integrates hardware and software. It can build a virtual model through augmented reality (Augmented Reality, AR), artificial intelligence (AI), 3D vision and other technologies, and generate a display based on the virtual model. picture. The virtual fitting system can realize the 360-degree natural fit of virtual clothes and the effect of clothes moving with people. It can be widely used in online shopping, daily wear and other life scenes.
虚拟试衣系统可以内置在智能终端、显示设备等智能设备中,在使用虚拟试衣系统时,控制智能设备运行虚拟试衣系统的应用程序。再输入照片等图像数据,则虚拟试衣系统可以通过图片合成的方式形成合成效果图,通过智能设备进行显示,达到虚拟试衣的目的。The virtual fitting system can be built into smart devices such as smart terminals and display devices. When using the virtual fitting system, the smart device is controlled to run the application of the virtual fitting system. Then input image data such as photos, then the virtual fitting system can form a synthetic effect map through image synthesis, and display it through smart devices to achieve the purpose of virtual fitting.
但是,这种通过图片合成的方式形成效果图像的虚拟试衣方式,只适用于静态的图像展示,无法实时展示试衣效果,也无法展示多种穿戴姿势、可视角度下的效果,导致这种虚拟试衣系统呈现的最终效果较差,试衣还原度低。However, this virtual fitting method, which forms effect images through image synthesis, is only suitable for static image display, and cannot display the fitting effect in real time, nor can it display the effects of various wearing postures and viewing angles. The final effect of this kind of virtual fitting system is poor, and the degree of restoration of fitting is low.
发明内容Contents of the invention
本公开提供一种显示设备,包括:显示器、摄像头、通信器以及控制器。其中,显示器被配置为显示图像和/或用户界面;摄像头被配置为实时采集图像数据,图像数据包括与用户关联的图像;通信器被配置为与服务器建立通信连接,服务器内置模型重建应用,用于根据与用户关联的图像数据生成人体模型;控制器与显示器、摄像头和通信器连接,控制器被配置为执行:获取与用户关联的图像数据;将与用户关联的图像数据发送给服务器,以使得服务器根据与用户关联的图像数据生成人体模型并向控制器发送人体模型;为人体模型添加衣物素材,以合成渲染模型;从所述与用户关联的图像数据中提取动作参数,以及按照所述动作参数调节所述渲染模型的模型姿态,以渲染出试衣画面。The present disclosure provides a display device, including: a display, a camera, a communicator, and a controller. Wherein, the display is configured to display images and/or user interfaces; the camera is configured to collect image data in real time, and the image data includes images associated with the user; the communicator is configured to establish a communication connection with the server, and the server has a built-in model reconstruction application, using The human body model is generated according to the image data associated with the user; the controller is connected with the display, the camera, and the communicator, and the controller is configured to perform: obtaining the image data associated with the user; sending the image data associated with the user to the server for Make the server generate a human body model according to the image data associated with the user and send the human body model to the controller; add clothing materials to the human body model to synthesize the rendering model; extract action parameters from the image data associated with the user, and according to the The action parameters adjust the model pose of the rendered model to render a fitting picture.
本公开还提供一种虚拟试衣系统,包括:显示设备、图像采集装置以及服务器;其中,图像采集装置连接显示设备,显示设备与服务器建立通信连接;图像采集装置被配置为实时采集图像数据,图像数据包括与用户关联的图像,以及对与用户关联的图像执行图像信号处理,以生成与用户关联的图像数据;图像采集装置还被配置为向显示设备发送与用户关联的图像数据;显示设备被配置为获取与用户关联的图像数据,以及将与用户关联的图像数据发送给服务器;服务器中内置模型重建应用,服务器被配置为接收与用户关联的图像数据,以及运行模型重建应用;根据与用户关联的图像数据生成人体模型,服务器还被配置为向显示设备发送人体模型;显示设备还被配置为向人体模型添加衣物素材,以合成渲染模型;以及,从与用户关联的图像数据中提取动作参数,并按照动作参数调节渲染模型的模型姿态,以渲染出试衣画面。The present disclosure also provides a virtual fitting system, including: a display device, an image acquisition device, and a server; wherein, the image acquisition device is connected to the display device, and the display device establishes a communication connection with the server; the image acquisition device is configured to collect image data in real time, The image data includes images associated with the user, and image signal processing is performed on the images associated with the user to generate image data associated with the user; the image acquisition device is also configured to send the image data associated with the user to the display device; the display device It is configured to obtain the image data associated with the user, and send the image data associated with the user to the server; a model reconstruction application is built in the server, and the server is configured to receive the image data associated with the user, and run the model reconstruction application; according to the The image data associated with the user generates a human body model, and the server is also configured to send the human body model to the display device; the display device is also configured to add clothing materials to the human body model to synthesize the rendered model; and, extract from the image data associated with the user Action parameters, and adjust the model pose of the rendered model according to the action parameters to render the fitting screen.
本公开还提供一种虚拟试衣方法,应用于虚拟试衣系统,虚拟试衣系统包括:显示设备、图像采集装置以及服务器;其中,图像采集装置连接显示设备,显示设备与服务器建立通信连接;虚拟试衣方法包括:图像采集装置实时采集与用户关联的图像,以及对与用户关联的图像执行图像信号处理,以生成与用户关联的图像数据;显示设备获取与用户关联的图像数据,以及将与用户关联的图像数据发送给服务器;服务器接收与用户关联的图像数据,根据与用户关联的图像数据生成人体模型,以及向显示设备发送人体模型;显示设备为人体模型添加衣物素材,以合成渲染模型;显示设备从与用户关联的图像数据中提取动作参数,并按照动作参数调节渲染模型的模型姿态,以渲染出试衣画面。The present disclosure also provides a virtual fitting method, which is applied to a virtual fitting system. The virtual fitting system includes: a display device, an image acquisition device, and a server; wherein, the image acquisition device is connected to the display device, and the display device establishes a communication connection with the server; The virtual fitting method includes: the image acquisition device collects images associated with the user in real time, and performs image signal processing on the image associated with the user to generate image data associated with the user; the display device obtains the image data associated with the user, and The image data associated with the user is sent to the server; the server receives the image data associated with the user, generates a mannequin based on the image data associated with the user, and sends the mannequin to the display device; the display device adds clothing materials to the mannequin for composite rendering Model; the display device extracts action parameters from the image data associated with the user, and adjusts the model pose of the rendering model according to the action parameters to render a fitting picture.
本公开实施例还提供一种计算机可读的非易失性存储介质,其上存储有计算机指令,所述计算机指令被处理器执行时使得计算机设备执行上述的方法。An embodiment of the present disclosure also provides a computer-readable non-volatile storage medium, on which computer instructions are stored, and when the computer instructions are executed by a processor, the computer device executes the above method.
附图说明Description of drawings
图1为本公开实施例中虚拟试衣系统使用场景图;FIG. 1 is a scene diagram of a virtual fitting system in an embodiment of the present disclosure;
图2为本公开实施例中显示设备的硬件配置图;FIG. 2 is a hardware configuration diagram of a display device in an embodiment of the present disclosure;
图3为本公开实施例中家用智能衣柜的结构示意图;FIG. 3 is a schematic structural diagram of a home smart wardrobe in an embodiment of the present disclosure;
图4为本公开实施例中内置摄像头的显示设备结构示意图;4 is a schematic structural diagram of a display device with a built-in camera in an embodiment of the present disclosure;
图5为本公开实施例中外接图像采集装置的显示设备结构示意图;5 is a schematic structural diagram of a display device connected to an external image acquisition device in an embodiment of the present disclosure;
图6a为本公开实施例中虚拟试衣界面示意图;Fig. 6a is a schematic diagram of a virtual fitting interface in an embodiment of the present disclosure;
图6b为本公开实施例中衣物选项分类显示效果图;Fig. 6b is a display effect diagram of clothing option classification in an embodiment of the present disclosure;
图6c为本公开实施例中识别衣物界面示意图;Fig. 6c is a schematic diagram of an interface for identifying clothes in an embodiment of the present disclosure;
图6d为本公开实施例中选择衣物颜色界面示意图;Fig. 6d is a schematic diagram of an interface for selecting a clothing color in an embodiment of the present disclosure;
图6e为本公开实施例中显示购买链接效果示意图;FIG. 6e is a schematic diagram showing the effect of purchase links in an embodiment of the present disclosure;
图6f为本公开实施例中显示购买界面示意图;FIG. 6f is a schematic diagram of a purchase interface displayed in an embodiment of the present disclosure;
图6g为本公开实施例中智能试衣镜显示商品销售位置效果示意图;Fig. 6g is a schematic diagram of the effect of displaying product sales positions in the smart fitting mirror in the embodiment of the present disclosure;
图7为本公开实施例中虚拟试衣方法流程示意图;7 is a schematic flow diagram of a virtual fitting method in an embodiment of the present disclosure;
图8为本公开实施例中服务器的软件配置图;FIG. 8 is a software configuration diagram of a server in an embodiment of the present disclosure;
图9为本公开实施例中一种试衣应用界面示意图;FIG. 9 is a schematic diagram of a fitting application interface in an embodiment of the present disclosure;
图10为本公开实施例中基于关键帧的动作驱动流程示意图;FIG. 10 is a schematic diagram of a key frame-based action driving process in an embodiment of the present disclosure;
图11为本公开实施例中匹配关联衣物流程示意图;Fig. 11 is a schematic diagram of a process flow for matching associated clothes in an embodiment of the present disclosure;
图12为本公开实施例中表情匹配流程示意图;FIG. 12 is a schematic diagram of an expression matching process in an embodiment of the present disclosure;
图13为本公开实施例中虚拟试衣系统数据交互时序图。Fig. 13 is a sequence diagram of data interaction in the virtual fitting system in an embodiment of the present disclosure.
具体实施方式Detailed ways
为使本申请的目的和实施方式更加清楚,下面将结合本申请示例性实施例中的附图,对本申请示例性实施方式进行清楚、完整地描述,显然,描述的示例性实施例仅是本申请一部分实施例,而不是全部的实施例。In order to make the purpose and implementation of the application clearer, the following will clearly and completely describe the exemplary implementation of the application in conjunction with the accompanying drawings in the exemplary embodiment of the application. Obviously, the described exemplary embodiment is only the present application. Claim some of the examples, not all of them.
需要说明的是,本申请中对于术语的简要说明,仅是为了方便理解接下来描述的实施方式,而不是意图限定本申请的实施方式。除非另有说明,这些术语应当按照其普通和通常的含义理解。It should be noted that the brief description of the terms in this application is only for the convenience of understanding the implementations described below, and is not intended to limit the implementations of this application. These terms are to be understood according to their ordinary and usual meaning unless otherwise stated.
本申请中说明书和权利要求书及上述附图中的术语“第一”、“第二”、“第三”等是用于区别类似或同类的对象或实体,而不必然意味着限定特定的顺序或先后次序,除非另外注明。应该理解这样使用的用语在适当情况下可以互换。The terms "first", "second", and "third" in the description and claims of this application and the above drawings are used to distinguish similar or similar objects or entities, and do not necessarily mean limiting specific sequential or sequential unless otherwise noted. It is to be understood that the terms so used are interchangeable under appropriate circumstances.
下面将详细地对实施例进行说明,其示例表示在附图中。下面的描述涉及附图时,除非另有表示,不同附图中的相同数字表示相同或相似的要素。以下实施例中描述的实施方式并不代表与本公开相一致的所有实施方式。仅是与权利要求书中所详述的、本公开的一些方面相一致的系统和方法的示例。The embodiments will be described in detail hereinafter, examples of which are illustrated in the accompanying drawings. When the following description refers to the accompanying drawings, the same numerals in different drawings refer to the same or similar elements unless otherwise indicated. The implementations described in the following examples do not represent all implementations consistent with this disclosure. These are merely examples of systems and methods consistent with aspects of the disclosure as recited in the claims.
在本公开实施例中,虚拟试衣系统是一种集成硬件和软件的服务平台,可以通过增强现实(Augmented Reality,AR)、人工智能(Artificial Intelligence,AI)以及3D视觉等技术,构建虚拟模型,并依据虚拟模型生成显示画面。In the embodiment of the present disclosure, the virtual fitting system is a service platform integrating hardware and software, which can construct a virtual model through augmented reality (Augmented Reality, AR), artificial intelligence (Artificial Intelligence, AI) and 3D vision technologies. , and generate a display screen based on the virtual model.
图1为本公开一些实施例的本申请实施例中虚拟试衣系统的示例性使用场景图。如图1所示,本申请本公开提供的虚拟试衣系统可以包括控制装置100、显示设备200、智能终端300、服务器400。虚拟试衣系统可以通过多个设备之间的协同工作,实现虚拟试衣功能。其中,控制装置100和智能终端300可以用于用户交互,可基于显示设备200提供虚拟试衣用户界面向显示设备200输入控制指令。显示设备200与服务器400均具有数据处理能力。显示设备200部署在本地,服务器400部署在云端,显示设备200与服务器400可以进行数据交互。Fig. 1 is an exemplary use scene diagram of the virtual fitting system in the embodiment of the present application of some embodiments of the present disclosure. As shown in FIG. 1 , the virtual fitting system provided by the present disclosure may include a control device 100 , a display device 200 , an intelligent terminal 300 , and a server 400 . The virtual fitting system can realize the virtual fitting function through the collaborative work among multiple devices. Wherein, the control device 100 and the smart terminal 300 can be used for user interaction, and can input control instructions to the display device 200 based on the virtual fitting user interface provided by the display device 200 . Both the display device 200 and the server 400 have data processing capability. The display device 200 is deployed locally, the server 400 is deployed in the cloud, and the display device 200 and the server 400 can exchange data.
控制装置100可接收用户的输入操作指令,且将操作指令转换为显示设备200可识别和响应的指令,起到用户与显示设备200之间交互中介作用。The control device 100 can receive the user's input operation instructions, and convert the operation instructions into instructions that the display device 200 can recognize and respond to, and play an intermediary role between the user and the display device 200 .
控制装置100可以是遥控器,遥控器和显示设备200的通信包括红外协议通信或蓝牙协议通信,及其他短距离通信方式,通过无线或有线方式来控制显示设备200。用户可以通过遥控器上按键、语音输入、控制面板输入等输入用户指令,来控制显示设备200。The control device 100 may be a remote controller, and the communication between the remote controller and the display device 200 includes infrared protocol communication, bluetooth protocol communication, and other short-distance communication methods, and the display device 200 is controlled wirelessly or wiredly. The user can control the display device 200 by inputting user commands through buttons on the remote control, voice input, control panel input, and the like.
在一些实施例中,也可以使用智能终端300(如移动终端、平板电脑、计算机、笔记本电脑等)以控制显示设备200。例如,使用在智能终端上运行的应用程序控制显示设备200。In some embodiments, a smart terminal 300 (such as a mobile terminal, a tablet computer, a computer, a notebook computer, etc.) can also be used to control the display device 200 . For example, the display device 200 is controlled using an application program running on a smart terminal.
在一些实施例中,显示设备可以不使用上述的智能终端或控制设备接收指令,而是通过触摸或者手势等接收用户的控制。In some embodiments, the display device may not use the above-mentioned smart terminal or control device to receive instructions, but may receive user control through touch or gesture.
在一些实施例中,显示设备200还可以采用除了控制装置100和智能终端300之外的方式进行控制,例如,可以通过显示设备200设备内部配置的获取语音指令的模块直接接收用户的语音指令控制,也可以通过显示设备200设备外部设置的语音控制设备来接收用户的语音指令控制。In some embodiments, the display device 200 can also be controlled in a manner other than the control device 100 and the smart terminal 300. For example, the module for acquiring voice commands configured inside the display device 200 can directly receive the user's voice command control , the user's voice command control can also be received through the voice control device installed outside the display device 200.
需要说明的是,显示设备200泛指具有数据处理能力和画面显示能力的设备,包括但不限于智能电视、个人计算机、显示终端、智能广告屏、魔方屏、虚拟现实设备、增强现实设备、智能可穿戴设备、机器人助手、商场的智能试衣镜、家庭智能衣柜等。在一些实施例中,图2是本公开一些实施例的显示设备的示例性硬件配置图。如图2所示,显示设备200包括供电电源210、通信器220、存储器230、接口模块240、控制器250、显示器260等功能模块中的一种或多种组合。It should be noted that the display device 200 generally refers to devices with data processing capabilities and screen display capabilities, including but not limited to smart TVs, personal computers, display terminals, smart advertising screens, Rubik's Cube screens, virtual reality devices, augmented reality devices, smart Wearable devices, robot assistants, smart fitting mirrors in shopping malls, smart wardrobes at home, etc. In some embodiments, FIG. 2 is an exemplary hardware configuration diagram of a display device in some embodiments of the present disclosure. As shown in FIG. 2 , the display device 200 includes one or more combinations of functional modules such as a power supply 210 , a communicator 220 , a memory 230 , an interface module 240 , a controller 250 , and a display 260 .
其中,供电电源210可以为显示设备200供电,使各功能模块可以上电运行。通信器220用于使显示设备200与服务器400建立通信连接关系,例如,显示设备200通过局域网(LAN)、无线局域网(WLAN)和其他网络进行通信连接。服务器400可以向显示设备200提供各种内容和互动。存储器230用于存储各种信息和应用程序数据。接口模块240用于使显示设备200连接外设,实现特定类型信号的输入或输出。控制器250通过运行存储在存储器230上中各种软件控制程序,来控制显示设备200的工作和响应用户的操作。Wherein, the power supply 210 can supply power to the display device 200 so that each functional module can be powered on and run. The communicator 220 is used to establish a communication connection relationship between the display device 200 and the server 400, for example, the display device 200 communicates through a local area network (LAN), a wireless local area network (WLAN) and other networks. The server 400 may provide various contents and interactions to the display device 200 . The memory 230 is used to store various information and application data. The interface module 240 is used to connect the display device 200 with peripheral devices to realize the input or output of specific types of signals. The controller 250 controls the operation of the display device 200 and responds to user's operations by running various software control programs stored in the memory 230 .
显示器260用于呈现用户界面,使显示设备200具有画面显示功能,显示设备200可以通过运行存储器230中与显示过程应用程序,在显示器260中呈现具体的显示画面。例如,播放界面、用户界面、应用程序界面等。The display 260 is used to present a user interface, so that the display device 200 has a screen display function, and the display device 200 can present a specific display screen on the display 260 by running the application program in the memory 230 and displaying the process. For example, playback interface, user interface, application program interface, etc.
对于不同类型的显示设备200,显示器260可以采用不同的形式并具有不同的显示比例。对于智能电视、个人计算机、显示终端等显示设备200,显示器260可以为标准显示画面比例对应的形状。例如,智能电视的显示分辨率为3840×2160;个人计算机的显示器分辨率为1920×1080;显示终端的显示分辨率为2400×1080。而对于虚拟现实设备、商场的智能试衣镜以及家庭智能衣柜等显示设备200,其所包含的显示器260可以按照实际用途设计相适应等的形状和比例。例如,虚拟现实设备的显示器260包括左右两个方形屏幕;图3示出本公开一些实施例的家用智能衣柜的结构示意图,如图3所示,家庭智能衣柜的显示器260为宽度和高度小于或等于衣柜柜门宽度和高度的长条形显示屏。For different types of display devices 200, the display 260 may take different forms and have different display ratios. For display devices 200 such as smart TVs, personal computers, and display terminals, the display 260 may have a shape corresponding to a standard display screen ratio. For example, the display resolution of a smart TV is 3840×2160; the display resolution of a personal computer is 1920×1080; the display resolution of a display terminal is 2400×1080. As for display devices 200 such as virtual reality devices, smart dressing mirrors in shopping malls, and home smart wardrobes, the displays 260 included in them can be designed in a shape and proportion suitable for actual purposes. For example, the display 260 of the virtual reality device includes two square screens on the left and right; FIG. 3 shows a schematic structural diagram of a home smart wardrobe in some embodiments of the present disclosure. As shown in FIG. Elongated display equal to the width and height of the wardrobe door.
由于虚拟试衣系统可以根据与用户关联的人像画面进行画面合成,从而生成试衣效果 画面。因此,为了能够进行画面合成,显示设备200应能够获取与用户关联的人像画面。在一些实施例中,显示设备200可以通过内置的图像采集模块270获取人像画面,即显示设备200在包括上述各种功能模块的基础上,还包括图像采集模块270。例如,图4示出本公开一些实施例的内置摄像头的显示设备的结构示意图,如图4所示,图像采集模块为设置在显示设备200顶部或底部的摄像头。Because the virtual fitting system can synthesize images according to the portrait images associated with the user, thereby generating a fitting effect image. Therefore, in order to perform frame composition, the display device 200 should be able to acquire a portrait frame associated with the user. In some embodiments, the display device 200 can acquire portrait images through the built-in image acquisition module 270 , that is, the display device 200 further includes the image acquisition module 270 on the basis of including the above-mentioned various functional modules. For example, FIG. 4 shows a schematic structural diagram of a display device with a built-in camera in some embodiments of the present disclosure. As shown in FIG. 4 , the image acquisition module is a camera disposed on the top or bottom of the display device 200 .
在一些实施例中,显示设备200可以通过外接图像采集装置500获取人像画面。图5示出本公开一些实施例的外接图像采集装置的显示设备结构示意图,即如图5所示,显示设备200可以通过接口模块240连接图像采集装置500。图像采集装置500内置摄像头和传输电路,可以通过摄像头对用户进行画面拍摄,再将拍摄获得的图像或视频通过传输电路和接口模块240发送给显示设备200进行显示。In some embodiments, the display device 200 can acquire portrait images through an external image acquisition device 500 . FIG. 5 shows a schematic structural diagram of a display device externally connected to an image acquisition device in some embodiments of the present disclosure, that is, as shown in FIG. 5 , the display device 200 can be connected to the image acquisition device 500 through the interface module 240 . The image acquisition device 500 has a built-in camera and a transmission circuit, and can take pictures of the user through the camera, and then send the captured image or video to the display device 200 for display through the transmission circuit and the interface module 240 .
在虚拟试衣系统中,显示设备200作为用户的直接交互设备,可以接收用户的控制指令,并根据控制指令执行数据处理,以形成包含不同内容的用户界面,通过显示器260进行呈现。在一些实施例中,显示设备200可以作为虚拟试衣系统的专用设备,即显示设备200仅用于运行虚拟试衣程序以及呈现虚拟试衣界面。例如,显示设备200可以应用于商场环境的机器人助手,机器人助手可以与用户进行语音交互,实现虚拟试衣功能。In the virtual fitting system, the display device 200 acts as a direct interaction device for the user, which can receive the user's control instructions, and perform data processing according to the control instructions to form a user interface containing different contents, which are presented through the display 260 . In some embodiments, the display device 200 can be used as a dedicated device of the virtual fitting system, that is, the display device 200 is only used to run the virtual fitting program and present the virtual fitting interface. For example, the display device 200 can be applied to a robot assistant in a shopping mall environment, and the robot assistant can perform voice interaction with the user to realize a virtual fitting function.
在一些实施例中,显示设备200还可以是虚拟试衣系统的一个实施设备,即显示设备200具有众多使用功能,虚拟试衣功能是众多使用功能中的一个。例如,显示设备200可以是个人计算机,可以在个人计算机中安装虚拟试衣应用,使其能够实施虚拟试衣功能。In some embodiments, the display device 200 can also be an implementation device of the virtual fitting system, that is, the display device 200 has many functions, and the virtual fitting function is one of the many functions. For example, the display device 200 may be a personal computer, and a virtual fitting application may be installed in the personal computer so that it can implement a virtual fitting function.
在一些实施例中,可以在显示设备200中安装多种应用程序,用于实现特定的功能。所安装的应用程序可以为系统应用或第三方应用。例如,显示设备200运营商提供的应用商店中可以展示供用户下载并安装的“虚拟试衣”应用程序。“虚拟试衣”应用程序中可以预先构建有衣物素材,则显示设备200响应用户的输入操作,可以运行该应用程序,接收与用户关联的图像,显示设备200可以对与用户关联的图像与预先构建的衣物素材进行合成,并显示合成效果,达到试衣目的。In some embodiments, various application programs can be installed in the display device 200 for realizing specific functions. The installed applications can be system applications or third-party applications. For example, a "virtual fitting" application program for users to download and install may be displayed in the application store provided by the operator of the display device 200 . Clothing material can be pre-built in the "virtual fitting" application program, then the display device 200 can run the application program in response to the user's input operation, and receive the image associated with the user, and the display device 200 can compare the image associated with the user with the pre-installed The constructed clothing material is synthesized, and the synthesis effect is displayed to achieve the purpose of fitting.
试衣类应用程序可以为独立的应用程序也可以是集成在特定应用中的一个功能模块。例如,对于购物类应用,为了使用户可以在网络购物过程中了解要购买衣物的穿戴效果,显示设备200可以在购物类应用中展示用于启动“虚拟试衣”功能的控件或图标。与独立应用相适应的,在启用“虚拟试衣”功能后,购物类应用可以通过向用户展示提示界面,该提示界面中包括用于引导用户输入与用户关联的图像的控件或图标。显示设备200调用当前带购买服装的素材模型,从而利用“虚拟试衣”功能将与用户关联的图像与素材模型进行合成,输出试衣效果画面。The fitting application program can be an independent application program or a functional module integrated in a specific application. For example, for a shopping application, in order to enable the user to understand the wearing effect of the clothing to be purchased during the online shopping process, the display device 200 may display a control or icon for activating the "virtual fitting" function in the shopping application. Compatible with independent applications, after enabling the "virtual fitting" function, the shopping application can display a prompt interface to the user, and the prompt interface includes controls or icons used to guide the user to input an image associated with the user. The display device 200 invokes the current material model with the purchased clothing, thereby using the "virtual fitting" function to synthesize the image associated with the user with the material model, and output a picture of the fitting effect.
显示设备200在实现虚拟试衣功能的过程中,可以基于获取的与用户关联的人像画面进行图像合成处理。在一些实施例中,显示设备200以人像画面为基础通过添加虚拟衣物图案,将人像画面与衣物画面进行合成,实现虚拟试衣。例如,显示设备200在获取与用户关联的人像画面后,可以对人像画面进行特征识别,识别出人像图案的可穿衣位置,包括上肢、下肢、手、足、颈部、头顶等。再从衣物素材库中提取虚拟衣物素材,从而在各可穿衣位置处添加对应的衣物素材,完成虚拟试衣。During the process of implementing the virtual fitting function, the display device 200 may perform image synthesis processing based on the acquired portrait picture associated with the user. In some embodiments, the display device 200 synthesizes the portrait image and the clothing image by adding virtual clothing patterns based on the portrait image to realize virtual fitting. For example, after acquiring the portrait picture associated with the user, the display device 200 can perform feature recognition on the portrait picture to identify the wearable position of the portrait pattern, including upper limbs, lower limbs, hands, feet, neck, top of the head, etc. Then extract the virtual clothing material from the clothing material library, so as to add the corresponding clothing material at each dressing position to complete the virtual fitting.
在一些实施例中,显示设备200还可以按照虚拟衣物图案为基础,通过添加人像图案进行画面合成。例如,虚拟试衣应用可以在虚拟衣物图案对应的头部区域添加人像显示区域,并将获取的人像图案画面中,头部画面显示在该显示区域内,从而合成出虚拟试衣画面。In some embodiments, the display device 200 may also perform screen synthesis by adding portrait patterns based on virtual clothing patterns. For example, the virtual fitting application can add a portrait display area to the head area corresponding to the virtual clothing pattern, and display the head image in the acquired portrait pattern picture in the display area, thereby synthesizing a virtual fitting picture.
其中,虚拟衣物图案可以通过存储在存储器230或云端服务器400中的衣物素材获得。例如,虚拟试衣系统的运营商可以根据当前销售的热门衣物进行多角度图像拍摄、3D建模等处理方式,生成虚拟衣物素材。生成的虚拟衣物素材可以存储在虚拟试衣系统的服务器400中,当显示设备200启用虚拟试衣应用,并选中待试穿的衣物时,显示设备200可以根据用户选择向服务器400请求虚拟衣物素材。Wherein, the virtual clothing pattern can be obtained through clothing materials stored in the memory 230 or the cloud server 400 . For example, the operator of the virtual fitting system can generate virtual clothing materials by performing multi-angle image shooting, 3D modeling and other processing methods according to the popular clothing currently on sale. The generated virtual clothing material can be stored in the server 400 of the virtual fitting system. When the display device 200 starts the virtual fitting application and selects the clothing to be tried on, the display device 200 can request the virtual clothing material from the server 400 according to the user's selection. .
在获取虚拟衣物素材后,显示设备200可以根据自身存储器230的使用情况,对请求的虚拟衣物素材进行缓存。因此,后续使用虚拟试衣系统时,根据用户选定的待试穿的衣物, 可以先在存储器230的本地缓存中先匹配对应虚拟衣物素材,当本地缓存中包括对应虚拟衣物素材时,提取该虚拟衣物素材。而当本地缓存中未匹配到对应虚拟衣物素材时,可以再次向服务器400请求对应的虚拟衣物素材。After obtaining the virtual clothing material, the display device 200 may cache the requested virtual clothing material according to the usage of its own memory 230 . Therefore, when using the virtual fitting system subsequently, according to the clothing to be tried on selected by the user, the corresponding virtual clothing material can be first matched in the local cache of the memory 230, and when the local cache includes the corresponding virtual clothing material, the corresponding virtual clothing material can be extracted Virtual clothing material. And when the corresponding virtual clothing material is not matched in the local cache, the corresponding virtual clothing material may be requested to the server 400 again.
在一些实施例中,虚拟衣物素材可以包括多个维度相互独立的参数,包括:衣物版型、颜色、材质等。显示设备200可以通过不同维度上参数间的排列组合,呈现出多种不同衣物样式。例如,对于相同版型的衣物可以采用相同的虚拟衣物素材,而通过在虚拟试穿时,调整颜色、材质等参数,获得不同衣物的外观效果。因此,显示设备200可以通过较少的模型数据组合出多种衣物素材,减少衣物模型的构建量以及请求衣物素材时的数据传输量。In some embodiments, the virtual clothing material may include parameters with multiple dimensions independent of each other, including: clothing style, color, material, and the like. The display device 200 can present a variety of different clothing styles through the arrangement and combination of parameters in different dimensions. For example, the same virtual clothing material can be used for the same type of clothing, and the appearance of different clothing can be obtained by adjusting parameters such as color and material during virtual try-on. Therefore, the display device 200 can combine multiple clothing materials with less model data, reducing the construction amount of clothing models and the amount of data transmission when requesting clothing materials.
为了合成出虚拟试衣效果,在一些实施例中,显示设备200可以在运行虚拟试衣应用后,为用户呈现虚拟试衣界面。图6a示出本公开一些实施例的虚拟试衣界面示意图,如图6a所示,在虚拟试衣界面中可以包括展示窗口、衣物选项、控制选项以及穿搭推荐选项等。虚拟试衣界面可以向用户展示不同的选项,显示设备200响应用户在虚拟试衣界面上的选择操作,控制合成虚拟试衣画面。例如,图像采集装置500采集的用户人像画面可以实时显示在展示窗口中,若多个衣物选项中被选中的为“衣物A”,则显示设备200可以从本地缓存或服务器400中调用“衣物A”对应的虚拟衣物素材,并通过展示窗口进行显示。同理,若多个衣物中被选中的为“衣物B”的选项,则显示设备200可以调用“衣物B”对应的虚拟衣物素材,并通过展示窗口进行显示。In order to synthesize a virtual fitting effect, in some embodiments, the display device 200 may present a virtual fitting interface for the user after running the virtual fitting application. Fig. 6a shows a schematic diagram of a virtual fitting interface of some embodiments of the present disclosure. As shown in Fig. 6a, the virtual fitting interface may include display windows, clothing options, control options, and outfit recommendation options. The virtual fitting interface can display different options to the user, and the display device 200 responds to the user's selection operation on the virtual fitting interface to control and synthesize a virtual fitting image. For example, the portrait image of the user collected by the image acquisition device 500 can be displayed in the display window in real time, and if the selected clothing option is "clothing A", the display device 200 can call the "clothing A" from the local cache or the server 400. ” and display the corresponding virtual clothing material through the display window. Similarly, if the option of "clothing B" is selected among the multiple clothes, the display device 200 may call the virtual clothes material corresponding to "clothes B" and display it through the display window.
虚拟试衣界面中的衣物选项可以按照实际穿戴位置进行分类,例如包括上装、下装、鞋、帽、饰品等分类类别。不用类别间的衣物选项支持同时选中,但同一类别中的衣物选项不支持多选。例如,图6b示出本公开一些实施例的示例性衣物选项分类显示效果图,如图6b所示,显示设备200响应于用户在选中一个上装中的“上衣A”选项的操作,以及选中一个下装中的“裤子B”的操作,则显示设备200可以在展示窗口中同时显示选中的上装衣物“上衣A”和下装衣物“裤子B”对应的画面。而对于同为上装的“上衣A”和“上衣B”,则不能被同时选中,而是通过后选中的衣物素材替代先选中的衣物素材。The clothing options in the virtual fitting interface can be classified according to the actual wearing position, such as tops, bottoms, shoes, hats, accessories and other categories. Clothing options between different categories support simultaneous selection, but clothing options in the same category do not support multiple selections. For example, FIG. 6b shows an exemplary clothing option classification display effect diagram of some embodiments of the present disclosure. As shown in FIG. For the operation of "pants B" in the bottoms, the display device 200 can simultaneously display images corresponding to the selected top clothing "top A" and bottom clothing "pants B" in the display window. As for "Top A" and "Top B" which are both tops, they cannot be selected at the same time, but the clothing material selected later replaces the clothing material selected first.
不同类别的虚拟衣物素材可采用不同的衣物版型,因此需要显示设备200调用不同的虚拟衣物素材。而在一些实施例中,由于衣物种类众多,因此相同类别的虚拟衣物素材还可以进一步划分为更细致的分类。例如,对于上装的分类类别可以按照穿戴季节进一步划分为春装、夏装、秋装、冬装等类别;也可以按照穿着位置进一步划分为外套、衬衫等类别。同理,对于进一步划分的不同类别,也需要显示设备200调用不同的虚拟衣物素材。Different types of virtual clothing materials can adopt different clothing styles, so the display device 200 needs to call different virtual clothing materials. However, in some embodiments, since there are many kinds of clothes, virtual clothes materials of the same category can be further divided into more detailed categories. For example, the classification of tops can be further divided into categories such as spring clothes, summer clothes, autumn clothes, and winter clothes according to the wearing season; it can also be further divided into categories such as coats and shirts according to the wearing position. Similarly, for different categories that are further divided, the display device 200 also needs to call different virtual clothing materials.
通过显示设备200进行虚拟试衣交互时,除通过上述衣物选项选择试穿衣物外,还可以通过其他交互方式选中试穿衣物。在一些实施例中,显示设备200通过获取用户输入的衣物商品名称、货号等信息来确定用户选择的待试穿的衣物。对于这类交互方式,显示设备200可以根据用户信息在衣物素材库中进行检索,以调用与输入信息相关的衣物素材。When performing virtual fitting interaction through the display device 200, in addition to selecting the clothes to try on through the above clothing options, the clothes to try on can also be selected through other interactive ways. In some embodiments, the display device 200 determines the clothing to be tried on selected by the user by acquiring information such as the product name and item number of the clothing input by the user. For this type of interaction, the display device 200 can search the clothing material library according to the user information, so as to call the clothing material related to the input information.
在一些实施例中,显示设备200还可以基于图像识别技术,对衣物的展示图片、实体画面、吊牌上的条形码等图像类信息进行识别,确定待试穿的衣物。例如,如图6a、图6c所示,图6c示出本公开一些实施例的识别衣物界面示意图,在虚拟试衣界面中可以包括“识别衣物”选项,若“识别衣物”选项为用户选择的选项,则显示设备200可以自动启动图像采集装置500获取衣物图像,并对衣物图像进行识别,并根据图像识别结果调用虚拟衣物素材。其中,对衣物图像进行识别的依据可以通过计算图像相似度的方式,计算拍摄获得的衣物图像与预设标准图像之间的相似度,将相似度最高的标准图像作为识别结果,并在数据库中调用该标准图像对应的虚拟衣物素材。In some embodiments, the display device 200 can also identify image information such as display pictures, physical images, and barcodes on hang tags based on image recognition technology to determine the clothing to be tried on. For example, as shown in Figure 6a and Figure 6c, Figure 6c shows a schematic diagram of an interface for identifying clothing in some embodiments of the present disclosure, and the virtual fitting interface may include the option "Identify Clothes", if the option "Identify Clothes" is selected by the user option, the display device 200 can automatically start the image acquisition device 500 to acquire the clothing image, and recognize the clothing image, and call the virtual clothing material according to the image recognition result. Among them, the basis for identifying the clothing image can be calculated by calculating the similarity between the captured clothing image and the preset standard image, and the standard image with the highest similarity is used as the recognition result, and stored in the database. Call the virtual clothing material corresponding to the standard image.
虚拟试衣界面中的一个衣物选项对应一类虚拟试衣素材。通过虚拟试衣界面确定用户选中的一个衣物选项后,可以通过控制选项调整衣物参数。例如,图6d示出本公开一些实施例的选择衣物颜色界面示意图,如图6d所示,在选中“衣物A”选项后,显示设备200可以在虚拟试衣界面中显示颜色控制选项,以便于向用户展示可以在预设的多个颜色控制选项中选择任一颜色的功能,控制展示窗口中所显示虚拟衣物画面的当前颜色。A clothing option in the virtual fitting interface corresponds to a type of virtual fitting material. After determining a clothing option selected by the user through the virtual fitting interface, the clothing parameters can be adjusted through the control options. For example, FIG. 6d shows a schematic diagram of an interface for selecting the color of clothing in some embodiments of the present disclosure. As shown in FIG. 6d, after selecting the "clothing A" option, the display device 200 can display color control options in the virtual fitting interface, so that Show the user the function of selecting any color from the preset multiple color control options, and control the current color of the virtual clothing screen displayed in the display window.
控制选项除用于控制选中衣物素材的颜色外,还可以用于控制调节展示窗口的显示画面,例如,旋转显示角度、局部放大、亮度调节、美颜功能等。基于用户选中的对应的选 项匹配的输入参数调节交互动作,控制显示设备200对展示窗口的呈现效果。In addition to controlling the color of the selected clothing material, the control options can also be used to control and adjust the display screen of the display window, for example, rotating the display angle, partially zooming in, adjusting brightness, and beautifying functions. The interactive action is adjusted based on the input parameters matched with the corresponding option selected by the user, and the presentation effect of the display window on the display device 200 is controlled.
穿搭推荐选项用于启用或关闭虚拟试衣应用的穿搭推荐功能。穿搭推荐功能可以在确定用户选中的任意衣物选项后,根据用户所选择的衣物,自动在展示窗口中显示推荐穿搭的衣物画面。穿搭推荐功能可以按照特定的风格算法,以用户选中的衣物选项为基准,在虚拟衣物素材库中匹配未选中分类类别的衣物素材。穿搭推荐算法可以基于颜色、类型、风格等维度进行统一匹配运算。例如,响应用户选中黑色、正装类型的上装的操作,穿搭推荐算法可以自动匹配黑色、正装类型的下装和鞋,并调用相应的虚拟衣物模型,与选中的上装虚拟衣物模型一同显示在展示窗口中。The outfit recommendation option is used to enable or disable the outfit recommendation function of the virtual fitting application. The clothing recommendation function can automatically display the recommended clothing screen in the display window according to the clothing selected by the user after determining any clothing options selected by the user. The clothing recommendation function can match the clothing materials of unselected categories in the virtual clothing material library according to a specific style algorithm and based on the clothing options selected by the user. The outfit recommendation algorithm can perform unified matching operations based on dimensions such as color, type, and style. For example, in response to the user's operation of selecting black and formal tops, the outfit recommendation algorithm can automatically match black and formal bottoms and shoes, and call the corresponding virtual clothing model, which will be displayed together with the selected top virtual clothing model in the display window.
除上述展示窗口、衣物选项、控制选项以及穿搭推荐选项等内容,在一些实施例中,显示设备200所呈现的虚拟试衣界面中还可以包括提示内容和链接选项。例如,虚拟试衣界面中还可以包括“保存穿搭”选项,当获取到用户通过选中多个衣物选项完成试衣的操作后,可以通过点击“保存穿搭”选项,保存当前展示窗口中所展示图案。同时,虚拟试衣应用还可以通过响应用户的点击“保存穿搭”的操作后,自动为用户展示对应衣物的购买链接或导购信息,以便用户可以根据虚拟试衣结果,购买相应的衣物。In addition to the display window, clothing options, control options, and outfit recommendation options, in some embodiments, the virtual fitting interface presented by the display device 200 may also include prompt content and link options. For example, the virtual fitting interface can also include a "save outfit" option. When it is obtained that the user has completed the fitting operation by selecting multiple clothing options, the user can click the "save outfit" option to save all items in the current display window. Showcase patterns. At the same time, the virtual fitting application can also automatically display the purchase link or shopping guide information of the corresponding clothing for the user after responding to the user's click "save outfit", so that the user can purchase the corresponding clothing according to the virtual fitting results.
显然,对于不同类型的显示设备200,在虚拟试衣界面所显示的提示内容和链接选项可以不同。例如,图6e示出本公开一些实施例的显示购买链接效果示意图,图6f示出本公开一些实施例的显示购买界面示意图,如图6e、图6f所示,在计算机运行购物类应用时,在获取到用户选中“保存穿搭”后,显示设备200可以呈现当前展示窗口中所显示虚拟衣物图案对应的商品链接,以供用户选择。在获取到用户选中任一商品链接后,显示设备200可以从虚拟试衣界面跳转到商品详情界面中。而在使用商场的智能试衣镜时,如果识别到用户选中“保存穿搭”,则可以在虚拟试衣界面中显示当前展示窗口中对应衣物的销售门店位置,如“衣物A可以在三楼A区306××购买到”,如图6g所示,图6g示出本公开一些实施例的智能试衣镜显示商品销售位置效果示意图。Obviously, for different types of display devices 200, the prompt content and link options displayed on the virtual fitting interface may be different. For example, FIG. 6e shows a schematic diagram of displaying purchase link effects in some embodiments of the present disclosure, and FIG. 6f shows a schematic diagram of displaying purchase interface in some embodiments of the present disclosure. As shown in FIG. 6e and FIG. 6f, when the computer runs shopping applications, After acquiring that the user selects "Save Outfit", the display device 200 may present a product link corresponding to the virtual clothing pattern displayed in the current display window for the user to select. After obtaining any product link selected by the user, the display device 200 can jump from the virtual fitting interface to the product detail interface. When using the smart fitting mirror in the shopping mall, if it is recognized that the user has selected "Save Outfit", the location of the sales store corresponding to the clothing in the current display window can be displayed on the virtual fitting interface, such as "Clothing A can be sold on the third floor. Area A 306 "purchased by ××", as shown in Figure 6g, which shows a schematic diagram of the effect of the smart fitting mirror displaying the product sales location in some embodiments of the present disclosure.
可见,无论虚拟试衣类应用还是购物类应用,为了输出试衣效果画面,都需要获取用户输入的与用户关联的图像。而由于不同的应用所呈现的最终试衣效果画面形式不同,因此可以通过不同的方式输入与用户关联的图像。即在一些实施例中,为了合成出2D、静态的试衣图像画面,应用向用户提示展示用于输入与用户关联的图像的图标或控件,响应用户的操作,打开文件管理器,并在特定的保存路径下选中图片文件,以该图片文件作为与用户关联的图像输入。显示设备200内置的摄像头或外接的图像采集装置还可以为用户进行拍照,并将拍照获得的图像作为与用户关联的图像输入应用。It can be seen that no matter the virtual fitting application or the shopping application, in order to output the fitting effect screen, it is necessary to obtain the image associated with the user input by the user. Since different applications present different forms of the final fitting effect screen, images associated with the user may be input in different ways. That is, in some embodiments, in order to synthesize a 2D, static fitting image screen, the application prompts the user to display an icon or control for inputting an image associated with the user, responds to the user's operation, opens the file manager, and selects Select a picture file in the save path of the file, and use the picture file as the image input associated with the user. The built-in camera of the display device 200 or the external image acquisition device can also take photos for the user, and use the images obtained by the photos as images associated with the user to input into the application.
但是,对于2D的静态的图像展示,无法实时展示试衣效果,也无法展示多种穿戴姿势、可视角度下的效果,导致这种虚拟试衣应用所呈现的最终效果较差。为此,在一些实施例中,虚拟试衣应用还可以提供动态的试衣图像画面,以便通过动态的试衣图像画面获取用户通过上传视频文件或实时录制视频文件的方式输入的与用户关联的图像。输入的视频图像可以在应用界面中的特定区域内进行展示,以获得更好的合成效果。However, for 2D static image display, it is impossible to display the fitting effect in real time, and it is also impossible to display the effects of various wearing postures and viewing angles, resulting in a poor final effect presented by this virtual fitting application. For this reason, in some embodiments, the virtual fitting application can also provide a dynamic fitting image screen, so that the dynamic fitting image screen can obtain the information associated with the user input by the user by uploading a video file or recording a video file in real time. image. The input video image can be displayed in a specific area in the application interface to obtain a better composite effect.
为了获得更好的试衣效果,在一些实施例中,显示设备200还可以通过增强现实技术,在用户上传的与用户关联的视频画面中添加虚拟衣物素材模型,并且通过对视频画面中的人物目标进行追踪,使衣物素材模型可以跟随视频中的人物目标产生动作变化。但是,这种动态试衣方式受限于AR技术的局限,导致衣物素材与视频中的人物目标的融合效果较差,衣物无法精确贴合人物目标。并且衣物跟随人物动作的展示效果存在较大延迟,因此虚拟试衣功能的交互体验较差。In order to obtain a better fitting effect, in some embodiments, the display device 200 can also use augmented reality technology to add a virtual clothing material model to the video screen uploaded by the user and associated with the user, and to The target is tracked, so that the clothing material model can follow the movement changes of the character target in the video. However, this dynamic fitting method is limited by the limitations of AR technology, resulting in poor fusion of clothing material and character targets in the video, and clothing cannot accurately fit the character targets. In addition, there is a large delay in the display effect of the clothes following the movements of the characters, so the interactive experience of the virtual fitting function is poor.
为了获得更好的虚拟试衣效果和交互体验,在本公开的部分实施例中提供一种虚拟试衣方法,方法可以应用于虚拟试衣系统,用于实时展示试衣效果画面。其中,虚拟试衣系统包括显示设备200、服务器400以及图像采集装置500。图像采集装置500用于实时采集与用户关联的图像,并经过图像信号处理,形成与用户关联的图像数据。服务器400中可以内置模型重建应用,模型重建应用可以根据与用户关联的图像数据生成人体模型。服务器400向显示设备200发送人体模型。显示设备200用于运行虚拟试衣应用程序,并对人体模型进行渲染,以合成并展示虚拟试衣画面。具体的,虚拟试衣方法包括以下内容:In order to obtain a better virtual fitting effect and interactive experience, some embodiments of the present disclosure provide a virtual fitting method, which can be applied to a virtual fitting system for displaying a fitting effect screen in real time. Wherein, the virtual fitting system includes a display device 200 , a server 400 and an image acquisition device 500 . The image collection device 500 is used to collect images associated with the user in real time, and process image signals to form image data associated with the user. A model reconstruction application can be built in the server 400, and the model reconstruction application can generate a human body model according to the image data associated with the user. The server 400 sends the human body model to the display device 200 . The display device 200 is used to run the virtual fitting application program, and render the human body model to synthesize and display the virtual fitting picture. Specifically, the virtual fitting method includes the following:
通过图像采集装置500实时采集与用户关联的图像。图像采集装置500可以包括摄像头和数据处理模块,其中,摄像头可以对用户所处环境进行图像拍摄,获得与用户关联的图像。数据处理模块可以对与用户关联的图像进行图像信号处理,以生成与用户关联的图像数据,并将与用户关联的图像数据作为一种图像信号输入显示设备200。The image associated with the user is collected in real time by the image collection device 500 . The image acquisition device 500 may include a camera and a data processing module, wherein the camera may capture images of the user's environment to obtain images associated with the user. The data processing module may perform image signal processing on the image associated with the user to generate image data associated with the user, and input the image data associated with the user into the display device 200 as an image signal.
在一些实施例中,图像采集装置500可以是内置在显示设备200中的功能模块,与显示设备200为一个整体设备。例如,如图4所示,图像采集装置500为显示设备200上的摄像头,可以受显示设备200中控制器250的统一控制,并且可以将采集到的与用户关联的图像直接发送给控制器250。显然,为了便于采集与用户关联的图像,摄像头应设置在显示设备200上的特定位置。例如,对于智能电视等显示设备200,摄像头可以设置在智能电视的顶部,并且摄像头的拍摄方向与智能电视的屏幕光发出方向相同,从而能够拍摄到位于屏幕前的与用户关联的图像。In some embodiments, the image acquisition device 500 may be a functional module built in the display device 200, and the display device 200 is an integral device. For example, as shown in FIG. 4, the image acquisition device 500 is a camera on the display device 200, which can be under the unified control of the controller 250 in the display device 200, and can directly send the collected images associated with the user to the controller 250. . Obviously, the camera should be set at a specific position on the display device 200 in order to collect images associated with the user. For example, for a display device 200 such as a smart TV, the camera can be set on the top of the smart TV, and the shooting direction of the camera is the same as the light emitting direction of the screen of the smart TV, so that images associated with the user located in front of the screen can be captured.
在一些实施例中,图像采集装置500还可以是外接显示设备200的外接设备,即如图5所示,显示设备200上可以设有接口模块240,接口模块240可以是高清晰度多媒体接口(High Definition Multimedia Interface,HDMI接口)、模拟或数据高清分量输入接口、复合视频广播信号(Composite Video Broadcast Signal,CVBS)接口、(通用串行总线(Universal Serial Bus,USB)接口以及RGB端口中的一种,可支持特定的数据传输方式。则图像采集装置500可以在接入接口模块240后,通过该接口所支持的特定数据传输方式,将采集到的与用户关联的图像发送给显示设备200。例如,图像采集装置500可以在连接智能电视后,通过开放自然交互(Open Natural Interaction,OpenNI)与智能电视进行通信,从而将采集到的与用户关联的图像发送给智能电视。In some embodiments, the image acquisition device 500 can also be an external device connected to the display device 200, that is, as shown in FIG. High Definition Multimedia Interface, HDMI interface), analog or data high-definition component input interface, composite video broadcast signal (Composite Video Broadcast Signal, CVBS) interface, (Universal Serial Bus (Universal Serial Bus, USB) interface and one of the RGB ports can support a specific data transmission method. Then the image acquisition device 500 can send the collected image associated with the user to the display device 200 through the specific data transmission method supported by the interface after accessing the interface module 240. For example, after being connected to the smart TV, the image acquisition device 500 may communicate with the smart TV through Open Natural Interaction (OpenNI), so as to send the collected image associated with the user to the smart TV.
为了实现更好的动态试衣效果,在图像采集装置500向显示设备200发送的与用户关联的图像数据中除包括图像内容外,还可以包括图像识别结果、骨骼参数、表情参数、手势识别结果数据等内容。因此,在一些实施例中,图像采集装置500的数据处理模块中还可以内置图像处理应用,当摄像头采集到与用户关联的图像后,数据处理模块可以通过运行图像处理应用,对与用户关联的图像进行识别。In order to achieve a better dynamic fitting effect, in addition to the image content, the image data associated with the user sent by the image acquisition device 500 to the display device 200 may also include image recognition results, bone parameters, expression parameters, and gesture recognition results. data etc. Therefore, in some embodiments, the data processing module of the image acquisition device 500 can also have a built-in image processing application. After the camera captures an image associated with the user, the data processing module can run the image processing application to process the image associated with the user. The image is recognized.
例如,图像采集装置500可以通过运行不同功能的图像处理应用,在与用户关联的图像中识别出“图像-深度”数据、3D人体骨骼关键点坐标、人像头部识别位置、人像目标追踪点等内容。从而在将识别出的内容发送给显示设备200时,可以为显示设备200提供RGBD图像数据、肢体驱动、人体跟踪、人脸重建素材以及表情驱动素材等功能支持。因此,这些数据可以与用户关联的图像一起构成与用户关联的图像数据。For example, the image acquisition device 500 can identify "image-depth" data, 3D human skeleton key point coordinates, portrait head recognition position, portrait target tracking point, etc. in the image associated with the user by running image processing applications with different functions. content. Therefore, when the recognized content is sent to the display device 200 , the display device 200 can be provided with functional supports such as RGBD image data, limb driving, human body tracking, face reconstruction material and expression driving material. Accordingly, these data may, together with user-associated images, constitute user-associated image data.
为了获得相应的数据,图像采集装置500需要具有特定的硬件支持。例如,为了能够在图像中识别出图像深度,图像采集装置500可以是具有多个镜头的摄像头组,多个镜头可以从不同位置检测同一目标,从而通过对比多个镜头的图像拍摄结果和角度差异,计算出目标的深度信息。In order to obtain corresponding data, the image acquisition device 500 needs to have specific hardware support. For example, in order to be able to identify the image depth in the image, the image acquisition device 500 can be a camera group with multiple lenses, and the multiple lenses can detect the same target from different positions, so that by comparing the image capture results and angle differences of multiple lenses , to calculate the depth information of the target.
显然,图像采集装置500还可以通过集成特定功能的传感器元件,使其采集的与用户关联的图像中包含用于进行三维建模的数据。例如,在图像采集装置500中,可以内置激光雷达元件,激光雷达可以向拍摄区域发射激光,并检测反射的激光信号测量拍摄区域内目标与图像采集装置500之间的距离,并将扫描结果与摄像头拍摄的每帧图像进行关联,从而生成带有图像深度的与用户关联的图像数据。Apparently, the image acquisition device 500 can also integrate sensor elements with specific functions, so that the images collected by the device 500 associated with the user include data for three-dimensional modeling. For example, in the image acquisition device 500, a laser radar component can be built in, and the laser radar can emit laser light to the shooting area, and detect the reflected laser signal to measure the distance between the target in the shooting area and the image acquisition device 500, and compare the scanning result with the image acquisition device 500. Each frame of image captured by the camera is correlated to generate user-associated image data with image depth.
如图7所示,图7示出本公开一些实施例的虚拟试衣方法流程示意图。图像采集装置500在采集与用户关联的图像并生成与用户关联的图像数据后,可以将与用户关联的图像数据发送给显示设备200。显示设备200则获取与用户关联的图像数据,作为试衣类应用的数据输入。显示设备200可以在试衣类应用启动后,向图像采集装置500发送启动指令,控制图像采集装置500开始采集与用户关联的图像,并对与用户关联的图像执行处理以获得与用户关联的图像数据。As shown in FIG. 7 , FIG. 7 shows a schematic flowchart of a virtual fitting method in some embodiments of the present disclosure. After the image acquisition device 500 acquires the image associated with the user and generates the image data associated with the user, it can send the image data associated with the user to the display device 200 . The display device 200 acquires the image data associated with the user as data input for the fitting application. The display device 200 may send an activation command to the image acquisition device 500 after the fitting application is started, control the image acquisition device 500 to start capturing images associated with the user, and perform processing on the images associated with the user to obtain images associated with the user data.
在一些实施例中,根据显示设备200所启动的试衣类应用所呈现的画面效果的不同,显示设备200可以向图像采集装置500获取包括不同内容的与用户关联的图像数据。例如,当显示设备200启动仅能够展示2D静态效果的试衣类应用时,显示设备200可以在向图像采 集装置500发送的启动指令中,设置与用户关联的图像数据的类型仅包括图像内容。而当显示设备200启动能够展示3D动态效果的试衣类应用时,则需要在启动指令中指定与用户关联的图像数据包括“图像-深度”数据、3D人体骨骼关键点坐标等数据内容。In some embodiments, according to the different image effects presented by the fitting application launched by the display device 200 , the display device 200 may acquire image data associated with the user including different contents from the image acquisition device 500 . For example, when the display device 200 starts a fitting application that can only display 2D static effects, the display device 200 may set the type of image data associated with the user to include only image content in the start instruction sent to the image capture device 500 . When the display device 200 starts a fitting application capable of displaying 3D dynamic effects, the image data associated with the user needs to be specified in the start command, including "image-depth" data, 3D human skeleton key point coordinates and other data content.
同理,对于不同的试衣类应用,显示设备200获取的与用户关联的图像数据中,图像的形式也可以不同。例如,对于静态效果的试衣类应用,可以从图像采集装置500中获取图片形式的与用户关联的图像数据;而对于动态效果的试衣类应用,则可以从图像采集装置500中获取视频形式的与用户关联的图像数据。Similarly, for different fitting applications, in the image data associated with the user acquired by the display device 200, the image forms may also be different. For example, for a fitting application with static effects, the image data associated with the user in the form of a picture can be obtained from the image acquisition device 500; image data associated with the user.
显示设备200在获取与用户关联的图像数据后,可以根据与用户关联的图像数据进行试衣结果画面的渲染。为了呈现更好的衣物融合效果,显示设备200可以基于人体模型获得试衣画面。因此在获取与用户关联的图像数据后,显示设备200可以根据与用户关联的图像数据创建人体模型。但是,由于人体模型的创建过程数据处理量巨大,并且受显示设备200硬件配置的限制,通过显示设备200本地创建的人体模型质量粗糙,精度较差。因此,在一些实施例中,显示设备200可以在获取与用户关联的图像数据后,将与用户关联的图像数据发送给服务器400进行建模处理。After the display device 200 acquires the image data associated with the user, it can render the fitting result screen according to the image data associated with the user. In order to present a better clothing fusion effect, the display device 200 may obtain a fitting screen based on a human body model. Therefore, after acquiring the image data associated with the user, the display device 200 can create a human body model according to the image data associated with the user. However, due to the huge amount of data processing during the creation of the human body model and the limitation of the hardware configuration of the display device 200, the quality of the human body model locally created by the display device 200 is rough and the accuracy is poor. Therefore, in some embodiments, after acquiring the image data associated with the user, the display device 200 may send the image data associated with the user to the server 400 for modeling processing.
为此,服务器400中可以内置模型重建应用。模型重建应用可以基于与用户关联的图像数据生成人体模型。在服务器400中,可以预先配置有多个初始人体模型,初始人体模型可以根据年龄、性别等因素设置多个。当显示设备200向服务器400发送与用户关联的图像数据后,服务器400可以先对与用户关联的图像数据中的图像进行识别,以确定图像中人物的年龄、性别等信息,再根据这些信息从预设的多个初始人体模型中调用相适应的一个。For this purpose, a model reconstruction application may be built in the server 400 . A model reconstruction application can generate a mannequin based on image data associated with a user. In the server 400, multiple initial human body models may be pre-configured, and multiple initial human body models may be set according to factors such as age and gender. After the display device 200 sends the image data associated with the user to the server 400, the server 400 can first identify the image in the image data associated with the user to determine the age, gender and other information of the person in the image, and then based on these information from Call the appropriate one among the preset multiple initial human models.
在调用初始人体模型后,服务器400还可以在与用户关联的图像数据中读取“图像-深度”数据、3D人体骨骼关键点坐标等数据,并根据读取的数据对初始人体模型进行设置、修改、调整,使初始模型逐渐具有与用户关联的图像数据中的人像特征。例如,服务器400可以从与用户关联的图像数据中提取用户头部宽度与肩部宽度之间的比例关系,并根据该比例关系调整初始模型中的头部宽度和肩部宽度。After invoking the initial human body model, the server 400 can also read data such as "image-depth" data and 3D human skeleton key point coordinates in the image data associated with the user, and set the initial human body model according to the read data, Modify, adjust, so that the initial model gradually has portrait features in the image data associated with the user. For example, the server 400 may extract the proportional relationship between the user's head width and shoulder width from the image data associated with the user, and adjust the head width and shoulder width in the initial model according to the proportional relationship.
在一些实施例中,如图8所示,图8示出本公开一些实施例的示例性服务器的软件配置图。服务器400所创建的人体模型可以包括多个部分,每个部分可以单独进行参数设置和调整,并且支持通过预设的素材库进行替换,以快速生成人体模型。例如,在服务器400中可以预先存储多种发型素材,服务器400在接收到与用户关联的图像数据后,可以根据与用户关联的图像数据中人像发型、颜色在发型素材库中匹配相类似的模型,并将头发模型添加到人体模型的头部,作为初始发型。In some embodiments, as shown in FIG. 8 , FIG. 8 shows a software configuration diagram of an exemplary server in some embodiments of the present disclosure. The human body model created by the server 400 may include multiple parts, each part can be parameterized and adjusted independently, and supports replacement through a preset material library, so as to quickly generate the human body model. For example, a variety of hairstyle materials can be pre-stored in the server 400. After the server 400 receives the image data associated with the user, it can match similar models in the hairstyle material library according to the portrait hairstyle and color in the image data associated with the user. , and add the hair model to the mannequin's head as the initial hairstyle.
而在后续使用中,试衣类应用具备可供用户任意选择更换虚拟试衣效果画面中的人物发型的功能,在获取到用户选择更换特定的发型后,显示设备200可以再向服务器400请求重建模型,则服务器400可以在发型素材库中选择新的发型素材添加到人体模型上,重建人体模型。In the subsequent use, the fitting application has the function of allowing the user to arbitrarily choose to change the hairstyle of the character in the virtual fitting effect screen. After obtaining the user's choice to change a specific hairstyle, the display device 200 can then request the server 400 to rebuild. model, the server 400 can select a new hairstyle material from the hairstyle material library and add it to the human body model to reconstruct the human body model.
为了创建更加精细的人体模型,在一些实施例中,可以在服务器400中内置多个建模模块,包括头部重建模块、身体重建模块、表情识别模块、试发模块等。每个建模模块可以通过特定的模型重建方式,构建对应的虚拟模型。例如,头部建模可以进一步包括几何重建和纹理重建单元,其中,几何重建单元可以根据与用户关联的图像数据进行点云去噪处理、三角构网处理、点云平滑处理以及点云融合处理,以使头部模型的形状与用户关联的图像数据中的人物目标逐渐趋于一致。而纹理重建单元可以对头部模型进行人像分割处理、肤色迁移处理以及肤色融合等处理,使头部模型的外观与用户关联的图像数据中的人物目标逐渐趋于一致。In order to create a more detailed human body model, in some embodiments, multiple modeling modules can be built in the server 400, including a head reconstruction module, a body reconstruction module, an expression recognition module, a trial hair module, and the like. Each modeling module can construct a corresponding virtual model through a specific model reconstruction method. For example, head modeling can further include geometric reconstruction and texture reconstruction units, wherein the geometric reconstruction unit can perform point cloud denoising processing, triangulation processing, point cloud smoothing processing, and point cloud fusion processing according to the image data associated with the user , so that the shape of the head model gradually tends to be consistent with the person target in the image data associated with the user. The texture reconstruction unit can perform portrait segmentation processing, skin color migration processing, and skin color fusion processing on the head model, so that the appearance of the head model is gradually consistent with the person target in the image data associated with the user.
同理,对于身体重建模块,也可以基于特定的建模方式生成身体部分的模型。例如,在身体重建模块可以通过蒙皮多人体线性模型(Skinned Multi-Person Linear Model,SMPL)方式创建身体模型。因此身体重建模块可以包括SMPL模型生成单元、参数化映射单元以及拼接单元等。在重建身体模型时,服务器400可以通过SMPL模型生成单元生成初始模型,再从与用户关联的图像数据中提取人体参数,以将SMPL模型向参数化模型进行参数映射, 得到符合与用户关联的图像数据的参数化身体模型,最后通过拼接单元将身体模型与头部模型进行拼接,得到人体模型。Similarly, for the body reconstruction module, models of body parts can also be generated based on specific modeling methods. For example, in the body reconstruction module, a body model can be created through a skinned multi-person linear model (Skinned Multi-Person Linear Model, SMPL). Therefore, the body reconstruction module may include an SMPL model generation unit, a parameterized mapping unit, and a stitching unit. When reconstructing the body model, the server 400 can generate an initial model through the SMPL model generation unit, and then extract human body parameters from the image data associated with the user, so as to perform parameter mapping from the SMPL model to the parameterized model, and obtain an image corresponding to the user. The parametric body model of the data, and finally the body model and the head model are spliced through the splicing unit to obtain the human body model.
在一些实施例中,为了得到人体模型,显示设备200还可以将与用户关联的图像数据发送给服务器400时,从与用户关联的图像数据中识别人像目标,并根据识别结果为人像目标添加骨骼关键点,以生成骨骼参数。再向服务器400发送骨骼参数,以使服务器400根据骨骼参数设置人体模型的关节点位置。添加骨骼参数后的人体模型可以按照骨骼参数对应的规律产生变化,从而能够模拟出更加符合真实的人物状态的模型姿态。In some embodiments, in order to obtain the human body model, when the display device 200 sends the image data associated with the user to the server 400, it can identify the portrait target from the image data associated with the user, and add bones to the portrait target according to the recognition result key points to generate bone parameters. Then send the skeleton parameters to the server 400, so that the server 400 can set the joint point positions of the human body model according to the skeleton parameters. The human body model after adding the bone parameters can change according to the rules corresponding to the bone parameters, so as to simulate a model posture that is more in line with the real state of the character.
此外,对于表情识别模块,也可以通过内置处理单元分别进行表情基模型建立以及表情参数估计,以根据与用户关联的图像数据中的人物表情为人体模型设置具体的表情形式。而对于试发模块,可以通过预置功能单元,分别用于设置头发匹配功能和头发穿透处理功能等,以创建更加真实的头发模型。In addition, for the expression recognition module, the built-in processing unit can also be used to establish the expression base model and estimate the expression parameters, so as to set specific expression forms for the human body model according to the expressions of the characters in the image data associated with the user. As for the trial hair module, the preset functional units can be used to set the hair matching function and hair penetration processing function, etc. to create a more realistic hair model.
由于服务器400生成或重建人体模型是用于后续的虚拟试衣过程,因此在一些实施例中,在服务器400中除内置有用于生成人体模型的应用外,还可以内置有用于生成衣物模型的应用。服务器400可以通过运行该应用,进行服装建模、衣料仿真以及服装变形的处理,以根据衣物图像数据创建衣物模型。Since the server 400 generates or reconstructs the human body model for the subsequent virtual fitting process, so in some embodiments, in addition to the built-in application for generating the human body model, the server 400 may also have a built-in application for generating the clothing model . The server 400 can perform clothing modeling, clothing material simulation, and clothing deformation processing by running the application, so as to create a clothing model based on clothing image data.
与人体模型相类似的,服务器400在创建衣物模型时,也可以由显示设备200向服务器400发送衣物图像数据,衣物图像数据可以通过对衣物进行多个角度的图像拍摄获得,例如,可以在衣物的商品界面中获取衣物的六视图,即正视图、背视图、左视图、右视图、仰视图以及俯视图。将不同角度的图片发送给服务器400后,服务器400可以根据这些图像自动建模生成衣服图案,再经过布料仿真、服装变形等渲染过程,获得符合真实效果的衣物模型。Similar to the human body model, when the server 400 creates the clothing model, the display device 200 can also send the clothing image data to the server 400. The clothing image data can be obtained by taking images of the clothing from multiple angles. Get six views of clothing in the product interface of , namely front view, back view, left view, right view, bottom view and top view. After sending the pictures from different angles to the server 400, the server 400 can automatically model and generate clothes patterns according to these images, and then through rendering processes such as cloth simulation and clothes deformation, a clothes model that conforms to the real effect can be obtained.
服务器400在生成人体模型后,可以向显示设备200发送生成的人体模型。显示设备200则在接收到人体模型后,对人体模型进行渲染,以渲染出试衣效果图像。在人体模型渲染的过程中,显示设备200可以将衣物素材添加至人体模型,以合成出渲染模型。After the server 400 generates the human body model, it may send the generated human body model to the display device 200 . After receiving the human body model, the display device 200 renders the human body model to render an image with a fitting effect. During the rendering of the human body model, the display device 200 may add clothing materials to the human body model to synthesize a rendered model.
在合成渲染模型的过程中,显示设备200还可以对服务器400发送的人体模型进行优化处理。例如,显示设备200可以基于Unity引擎对人体模型进行头部合成和还原处理,包括通过meshfilter工具和meshrender工具进行头部形状处理;通过texture工具和normalmap工具进行纹理处理;通过blendshape/morph工具进行头部参数模型算法进行表情参数信息处理和头发贴合头部形状的调整等处理过程。During the process of synthesizing the rendering model, the display device 200 may also optimize the human body model sent by the server 400 . For example, the display device 200 can perform head synthesis and restoration processing on the human body model based on the Unity engine, including performing head shape processing through the meshfilter tool and meshrender tool; performing texture processing through the texture tool and normalmap tool; and performing head shape processing through the blendshape/morph tool. The face parameter model algorithm is used to process expression parameter information and adjust the hair to fit the head shape.
在一些实施例中,为了便于向用户展示对比试衣效果,显示设备200在将与用户关联的图像数据发送给服务器400的同时,还可以对与用户关联的图像数据中的图像部分进行显示。例如,如图9所示,图9示出本公开一些实施例的试衣应用界面示意图。在试衣类应用程序的程序界面中,可以包括试衣窗口和原始窗口,在试衣窗口中可以显示人体模型及衣物素材的渲染结果,在原始窗口中则显示与用户关联的图像数据中的图像部分,即图像采集装置500实时拍摄的图像画面。In some embodiments, in order to show the user the effect of comparative fitting, the display device 200 may also display the image part in the image data associated with the user while sending the image data associated with the user to the server 400 . For example, as shown in FIG. 9 , FIG. 9 shows a schematic diagram of a fitting application interface of some embodiments of the present disclosure. The program interface of a fitting application program can include a fitting window and an original window. The rendering results of the human body model and clothing materials can be displayed in the fitting window, and the image data associated with the user can be displayed in the original window. The image part is the image frame captured by the image acquisition device 500 in real time.
显示设备200在合成出渲染模型后,还可以从与用户关联的图像数据中提取动作参数,并按照动作参数调节渲染模型的模型姿态,以渲染出试衣画面。即显示设备200可以根据与用户关联的图像数据确定用户的动作,并控制渲染模型跟随与用户关联的图像数据对应的用户动作。其中,可以在试衣画面中呈现的动作包括身体动作、头部动作、表情动作、手势动作以及衣物动作等。动作参数可以由图像采集装置500通过头部检测、3D人体骨架关键点检测、面部表情检测以及手势识别检测等方式计算获得。After the display device 200 synthesizes the rendered model, it can also extract action parameters from the image data associated with the user, and adjust the model pose of the rendered model according to the action parameters to render a fitting picture. That is, the display device 200 may determine the user's action according to the image data associated with the user, and control the rendering model to follow the user's action corresponding to the image data associated with the user. Among them, the actions that can be presented in the fitting screen include body actions, head actions, facial expressions, gesture actions, and clothing actions. The motion parameters can be calculated and obtained by the image acquisition device 500 through head detection, 3D human skeleton key point detection, facial expression detection, and gesture recognition detection.
显示设备200可以通过与用户关联的图像数据中多帧图像进行对比的方式,检测用户的动作。即在一些实施例中,显示设备200可以在按照动作参数调节渲染模型的模型姿态的步骤中,遍历与用户关联的图像数据中每帧图像的骨骼关键点;再对比相邻两帧图像中的骨骼关键点位置,以获得每个骨骼关键点的移动距离,以根据移动距离移动人体模型的关节点位置。通过对多个骨骼关键点的协同调整,显示设备200可以将渲染模型中的人物姿态跟随与用户关联的图像数据中的每帧图像内容调整为相同的动作,实现动作跟随效果。The display device 200 can detect the user's action by comparing multiple frames of images in the image data associated with the user. That is, in some embodiments, in the step of adjusting the model pose of the rendered model according to the action parameters, the display device 200 can traverse the skeleton key points of each frame image in the image data associated with the user; Skeleton key point position to obtain the movement distance of each bone key point to move the joint point position of the human body model according to the movement distance. Through the coordinated adjustment of multiple skeleton key points, the display device 200 can adjust the pose of the person in the rendered model to follow the image content of each frame of the image data associated with the user to the same action, realizing the action following effect.
为了实现动作跟随效果,显示设备200可以通过蒙皮,即为模型添加骨骼的方式使人 体模型产生合理的运行方式。例如,在获取人体模型后,显示设备200可以蒙皮到人体模型、衣物模型以及为布料仿真算法提供个纹理进行蒙皮,以及使用blendshape/morph工具为模型处理骨骼蒙皮动画。In order to achieve the action following effect, the display device 200 can make the human body model produce a reasonable running mode by skinning, that is, adding bones to the model. For example, after obtaining the human body model, the display device 200 can skin the human body model and the clothing model and provide a texture for the cloth simulation algorithm for skinning, and use the blendshape/morph tool to process the skeleton skinning animation for the model.
可见,在上述实施例中,虚拟试衣方法可以通过图像采集装置500实时采集与用户关联的图像数据,并通过服务器400根据与用户关联的图像数据生成人体模型,以及在显示设备200中为人体模型添加衣物素材,形成渲染模型。同时,显示设备200还可以根据与用户关联的图像数据获取用户动作,并根据用户动作实时调整渲染模型的模型姿态,实现虚拟形象驱动。方法可以通过人体模型,实现3D、动态、实时、多角度显示试衣画面的效果。并通过服务器400执行模型建立和重建,能够分担显示设备200的数据处理量,有利于生成更加精细、逼真的人物模型。It can be seen that, in the above-mentioned embodiment, the virtual fitting method can collect image data associated with the user through the image acquisition device 500 in real time, and generate a human body model through the server 400 according to the image data associated with the user, and display a human body model on the display device 200. Add clothing materials to the model to form a rendered model. At the same time, the display device 200 can also acquire user actions according to the image data associated with the user, and adjust the model posture of the rendering model in real time according to the user actions, so as to realize avatar driving. The method can realize the effect of 3D, dynamic, real-time and multi-angle display of the fitting screen through the human body model. Moreover, the server 400 executes model building and reconstruction, which can share the data processing load of the display device 200 and is conducive to generating a more detailed and realistic character model.
由于显示设备200部署在本地,服务器400部署在云端,显示设备200和服务器400之间进行数据通信时会受到网络延迟等因素的影响,导致在将与用户关联的图像数据发送给服务器400后,可能延迟一段时间才能后接收到人体模型。延长了展示试衣效果的等待时间,因此,在呈现最终试衣画面的过程中,应尽量减少显示设备200与服务器400之间的数据交互次数,以降低延迟。对此,在一些实施例中,显示设备200可以在初次获得人体模型后,对人像目标进行监控,从而在未发生较明显的特征变化时,一直使用初次获得的人体模型。即如图10所示,图10示出本公开一些实施例的基于关键帧的动作驱动流程示意图。显示设备200在将与用户关联的图像数据发送给服务器400后,图10至少包括如下步骤:Since the display device 200 is deployed locally and the server 400 is deployed in the cloud, the data communication between the display device 200 and the server 400 will be affected by factors such as network delay, resulting in that after the image data associated with the user is sent to the server 400, There may be a delay in receiving the manikin. The waiting time for displaying the fitting effect is prolonged. Therefore, in the process of presenting the final fitting screen, the number of data interactions between the display device 200 and the server 400 should be minimized to reduce delay. For this, in some embodiments, the display device 200 may monitor the portrait target after obtaining the human body model for the first time, so that the human body model obtained for the first time is always used when there is no obvious characteristic change. That is, as shown in FIG. 10 , FIG. 10 shows a schematic diagram of a key frame-based action driving process in some embodiments of the present disclosure. After the display device 200 sends the image data associated with the user to the server 400, FIG. 10 includes at least the following steps:
S1001、从与用户关联的图像数据中提取初始关键帧图像。S1001. Extract an initial key frame image from image data associated with a user.
S1002、在初始关键帧图像中提取图像深度参数,记录初始关键帧图像。S1002. Extract image depth parameters from the initial key frame image, and record the initial key frame image.
S1003、将初始关键帧和图像深度参数均发送给服务器400。S1003. Send both the initial key frame and the image depth parameter to the server 400.
服务器400根据图像深度参数设置人体模型参数,向显示设备200发送人体模型。显示设备200在获得人体模型后,显示设备200可以在后续应用中,执行S1004。The server 400 sets the human body model parameters according to the image depth parameters, and sends the human body model to the display device 200 . After the display device 200 obtains the human body model, the display device 200 may execute S1004 in subsequent applications.
S1004、从与用户关联的图像数据中提取实时关键帧图像。S1004. Extract real-time key frame images from the image data associated with the user.
其中,关键帧可以是特定时间点对应的一帧图像。例如,显示设备200可以每隔1s从与用户关联的图像数据中提取一帧图像作为关键帧。关键也可以是通过间隔一定帧数,获得的一帧图像,例如,显示设备200可以间隔20帧从与用户关联的图像数据中提取一帧图像作为关键帧。关键帧还可以是通过对图像帧进行图像识别,获得的带有特征明显的人像目标的一帧图像。例如,通过将与用户关联的图像数据逐帧输入人像识别模型中,则通过模型识别确定图像帧中包含人像时,将其标记为关键帧。Wherein, the key frame may be a frame of image corresponding to a specific time point. For example, the display device 200 may extract a frame of image from the image data associated with the user every 1 second as a key frame. The key may also be a frame of image obtained at intervals of a certain number of frames. For example, the display device 200 may extract a frame of image from image data associated with the user at an interval of 20 frames as a key frame. The key frame may also be a frame of image with a distinctive portrait target obtained by performing image recognition on the image frame. For example, by inputting the image data associated with the user into the portrait recognition model frame by frame, when it is determined by the model recognition that the image frame contains a portrait, it will be marked as a key frame.
S1005、对比初始关键帧图像和实时关键帧图像,以获得图像相似度。S1005. Comparing the initial key frame image and the real-time key frame image to obtain image similarity.
其中,初始关键帧为用于提取图像深度参数,以生成人体模型的图像帧。实时关键帧则是显示设备200在与用户关联的图像数据中实时提取的关键帧。图像相似度可以通过分别计算两幅图像的直方图,再计算两个直方图的归一化相关系数,如巴氏距离,直方图相交距离等数据,确定两个图像的相似度。由于用户处于静止状态时,两个关键帧图像的内容相同,而当用户处于动作状态时,两个关键帧图像的内容则存在差异,并且动作幅度越大,两帧图像的内容相似程度越低,因此可以通过图像相似度检测用户的动作状态。Wherein, the initial key frame is an image frame used to extract image depth parameters to generate a human body model. The real-time key frame is the key frame extracted by the display device 200 in real time from the image data associated with the user. The image similarity can be determined by calculating the histograms of the two images separately, and then calculating the normalized correlation coefficient of the two histograms, such as Bhattacharyachian distance, histogram intersection distance and other data, to determine the similarity of the two images. When the user is in a static state, the content of the two key frame images is the same, but when the user is in an action state, the content of the two key frame images is different, and the greater the range of motion, the lower the similarity of the content of the two frame images , so the user's action state can be detected by image similarity.
S1006、判断图像相似度是否大于预设相似度阈值,若是则执行S1007,否则执行S1008。S1006. Determine whether the image similarity is greater than a preset similarity threshold, if so, execute S1007, otherwise execute S1008.
在对比初始关键帧和实时关键帧图像以获得图像相似度后,显示设备200可以将图像相似度与预设相似度阈值进行对比。After comparing the initial key frame and real-time key frame images to obtain the image similarity, the display device 200 may compare the image similarity with a preset similarity threshold.
S1007、提取动作参数。S1007. Extracting action parameters.
如果图像相似度大于或等于预设相似度阈值,即两个关键帧中用户的实际动作幅度较小,或处于静止状态,此时无需重新建模,仍然采用原有人体模型即可获得较好的试衣画面展示效果,即显示设备200可以执行从与用户关联的图像数据中提取动作参数的步骤,以根据动作参数驱动人体模型产生动作。If the image similarity is greater than or equal to the preset similarity threshold, that is, the actual motion range of the user in the two key frames is small, or in a static state, there is no need to re-model at this time, and the original human body model can still be used to obtain better results. The display effect of the fitting screen, that is, the display device 200 can execute the step of extracting motion parameters from the image data associated with the user, so as to drive the human body model to generate motions according to the motion parameters.
S1008、从实时关键帧图像中提取图像深度参数。S1008. Extract image depth parameters from real-time key frame images.
如果图像相似度小于预设相似度阈值,即两个关键帧中的内容差异较大,此时为了获得更加逼真的模型效果,显示设备200可以从实时关键帧图像中提取图像深度参数,并将 提取的图像深度参数发送给服务器400,以使得服务器400根据图像深度参数重建人体模型。If the image similarity is less than the preset similarity threshold, that is, the content in the two key frames differs greatly, at this time, in order to obtain a more realistic model effect, the display device 200 can extract the image depth parameter from the real-time key frame image, and The extracted image depth parameters are sent to the server 400, so that the server 400 reconstructs the human body model according to the image depth parameters.
随着图像采集装置500获取持续获取与用户关联的图像,图像中所包含的画面内容可以随用户动作逐渐出现差异,即初始关键帧与后续提取的实施关键帧中图像相似度逐渐降低,因此,为了保持显示设备200对模型动作驱动的连贯性,减少重新建模次数。As the image acquisition device 500 continuously acquires images associated with the user, the screen content included in the image may gradually vary with the user's actions, that is, the image similarity between the initial key frame and the subsequently extracted key frame gradually decreases. Therefore, In order to maintain the consistency of the display device 200 driving the model action, reduce the number of times of remodeling.
S1009、使用实时关键帧图像替换记录的初始关键帧图像。S1009. Replace the recorded initial key frame image with the real-time key frame image.
在一些实施例中,显示设备200可以在初始关键帧图像中提取图像深度参数后,记录初始关键帧图像。并且在执行从与用户关联的图像数据中提取动作参数的步骤后,或者在将图像深度参数发送给服务器的步骤后,使用实时关键帧图像替换记录的初始关键帧图像。In some embodiments, the display device 200 may record the initial key frame image after extracting the image depth parameter from the initial key frame image. And after the step of extracting motion parameters from the image data associated with the user is performed, or after the step of sending the image depth parameters to the server, the real-time key frame images are used to replace the recorded initial key frame images.
例如,在试衣类应用启动后,显示设备200可以先从与用户关联的图像数据中提取初始关键帧T0,并将初始关键帧T0发送给服务器400生成人体模型。并且,显示设备200在后续渲染人体模型试衣画面时,可以持续获取实时关键帧,即T1。则显示设备200可以在获取实时关键帧T1后,可以对比初始关键帧T0和实时关键帧T1,计算两帧图像的相似度S01,当相似度S01大于或等于相似度阈值S时,则从实时关键帧T1或与用户关联的图像数据中提取动作参数,以驱动人体模型产生动作,跟随用户动作变化。在提取动作参数后,显示设备200可以使用实时关键帧T1替换初始关键帧T0,作为后续动作判断过程的初始关键帧。即在获得实时关键帧T2时,可以对比关键帧T1和关键帧T2之间的相似度,以继续追踪用户动作或者重建人体模型。For example, after a fitting application is started, the display device 200 may first extract an initial key frame T0 from image data associated with the user, and send the initial key frame T0 to the server 400 to generate a human body model. Moreover, when the display device 200 subsequently renders the fitting picture of the mannequin, it can continuously acquire the real-time key frame, that is, T1. Then the display device 200 can compare the initial key frame T0 and the real-time key frame T1 after acquiring the real-time key frame T1, and calculate the similarity S01 of the two frames of images. When the similarity S01 is greater than or equal to the similarity threshold S, then start Action parameters are extracted from the key frame T1 or the image data associated with the user to drive the human body model to generate actions and follow the changes of the user's actions. After extracting the motion parameters, the display device 200 may use the real-time key frame T1 to replace the initial key frame T0 as the initial key frame in the subsequent motion judgment process. That is, when the real-time key frame T2 is obtained, the similarity between the key frame T1 and the key frame T2 can be compared to continue tracking user actions or reconstructing the human body model.
有此可知,上述实施例可以通过初始关键帧和实时关键帧对图像采集装置500采集到的与用户关联的图像画面进行持续分析,从而追踪用户的实际动作。并且,显示设备200可以通过对比相邻两个关键帧之间的图像相似度,确定是否需要重建人体模型,以使人体模型在能够及时同步的前提下,减少重建模型的次数,提高试衣类应用的响应速度。It can be seen that, the above embodiment can continuously analyze the image frame associated with the user collected by the image collection device 500 through the initial key frame and the real-time key frame, so as to track the actual action of the user. Moreover, the display device 200 can determine whether it is necessary to reconstruct the human body model by comparing the image similarity between two adjacent key frames, so that the human body model can reduce the number of reconstruction models on the premise that it can be synchronized in time, and improve the fitting class. The responsiveness of the application.
在一些实施例中,试衣类应用在提供实时动态试衣功能的同时,还可以为用户提供穿搭推荐,即如图11所示,图11示出本公开一些实施例的匹配关联衣物流程示意图。显示设备200在为人体模型添加衣物素材的过程中,可以获取用户输入的用于选定衣物的选择指令。例如,试衣类应用界面中,可以设置有多个衣物选项,可以通过获取用户的交互操作确定选中的界面中的任一选项,以输入选择指令。还可以识别用户在多个衣物选项中,选中多个目标衣物,即选择指令中指定至少一个目标衣物。例如,选择指令可以是用户可以同时选中上装、下装以输入选择指令。In some embodiments, while the fitting application provides real-time dynamic fitting functions, it can also provide users with outfit recommendations, as shown in Figure 11, which shows the process of matching associated clothes in some embodiments of the present disclosure schematic diagram. During the process of adding clothing material to the human body model, the display device 200 may acquire a selection instruction input by the user for selecting clothing. For example, in a fitting application interface, multiple clothing options can be set, and any option in the interface that is selected can be determined by obtaining user interaction operations, so as to input a selection instruction. It can also be recognized that the user selects multiple target clothes among multiple clothes options, that is, at least one target clothes is specified in the selection command. For example, the selection instruction may be that the user can simultaneously select upload and download to input the selection instruction.
在显示设备200接收到用户的输入选择指令后,显示设备200可以响应于该选择指令,从衣物素材库中提取目标衣物素材。由于衣物可以划分为多个种类,如上装、下装、鞋、帽、包等。这些种类可以相互搭配,形成最终的穿衣效果。因此,当选择指令表征没有选择全部种类时,试衣类应用可以根据已选择的衣物特点,自动搭配相适应的其余类型服装,以便于呈现更好的试衣效果。After the display device 200 receives the user's input selection instruction, the display device 200 may respond to the selection instruction and extract the target clothing material from the clothing material library. Because clothing can be divided into multiple categories, such as tops, bottoms, shoes, hats, bags, etc. These types can be matched with each other to form the final dressing effect. Therefore, when the selection instruction indicates that not all types are selected, the fitting application can automatically match other suitable types of clothing according to the characteristics of the selected clothing, so as to present a better fitting effect.
因此,显示设备200在提取目标衣物素材的同时,可以按照预设穿搭推荐规则,根据目标衣物素材在衣物素材库中匹配关联衣物素材。其中,预设穿搭推荐规则可以基于颜色、用途、版型、适用年龄等分类综合设定。例如,当用户的选择为蓝白色小清新风格的上装时,可以按照预设穿搭推荐规则为用户推荐蓝白色小清新风格的下装、鞋等关联衣物素材。Therefore, while extracting the target clothing material, the display device 200 can match the associated clothing material in the clothing material library according to the target clothing material according to the preset dressing recommendation rules. Among them, the preset recommendation rules for outfits can be comprehensively set based on categories such as color, purpose, style, and applicable age. For example, when the user chooses blue and white tops with a small and fresh style, related clothing materials such as blue and white bottoms and shoes can be recommended to the user according to the preset recommendation rules for outfitting.
在获得目标衣物素材和关联衣物素材后,显示设备200可以向人体模型添加目标衣物素材和关联衣物素材,以形成带有意见穿搭效果的渲染模型。显然,为了给用户推荐更多类型的穿搭风格,显示设备200中可以预先设有多个穿搭推荐规则,每种规则可以根据用户已选择的目标衣物素材,匹配该规则下合适的关联衣物素材,并通过多个窗口进行预览,以提高显示效果。After obtaining the target clothing material and the associated clothing material, the display device 200 may add the target clothing material and the associated clothing material to the human body model, so as to form a rendering model with the effect of dressing in the opinion. Obviously, in order to recommend more types of dressing styles to the user, the display device 200 can pre-set multiple dressing recommendation rules, and each rule can match the appropriate associated clothing under the rule according to the target clothing material selected by the user. footage and preview it through multiple windows for enhanced display.
由于用户的实际表情的复杂程度较高,在不同的表情下人体模型需要修改的参数数据量巨大,导致试衣类应用在进行表情追踪时,需要处理大量的数据,以修改人体模型中的面部特征。这将导致应用的响应速度变慢,因此,在一些实施例中,显示设备200可以通过预设标准表情模板的形式,快速实现表情切换。即如图12所示,图12示出本公开一些实施例的表情匹配流程示意图。在按照动作参数调节渲染模型的模型姿态时,显示设备200可以在与用户关联的图像数据中识别头部区域,并在头部区域中检测用户的表情,再根据 用户的表情所属的表情类型,在预设表情库中匹配与表情类型相同的表情模型,从而使用表情模型替换渲染模型中的面部区域。Due to the high complexity of the user's actual expression, the amount of parameter data that needs to be modified on the human body model under different expressions is huge. As a result, fitting applications need to process a large amount of data when performing expression tracking to modify the face in the human body model. feature. This will slow down the response speed of the application. Therefore, in some embodiments, the display device 200 can quickly implement expression switching in the form of a preset standard expression template. That is, as shown in FIG. 12 , FIG. 12 shows a schematic diagram of an expression matching process in some embodiments of the present disclosure. When adjusting the model pose of the rendering model according to the action parameters, the display device 200 can identify the head area in the image data associated with the user, and detect the user's expression in the head area, and then according to the expression type of the user's expression, Match the expression model of the same type as the expression in the preset expression library, thereby replacing the facial area in the rendered model with the expression model.
例如,显示设备200可以通过检测与用户关联的图像数据中目标形状,以及该形状中画面布局特点,识别与用户关联的图像中的头部区域。再从头部区域中,使用表情识别模型识别当前用户的表情。其中,表情识别模型可以通过样本数据训练人工智能模型的方式获得。即通过将大量带有表情标签的样本数据输入初始模型,并设置模型输出结果为图像归属于特定表情类别的分类概率,再通过计算分类概率与表情标签的误差,并按照该误差反向传播调整模型参数,以使模型输出的结果逐渐与标签结果相同,得到表情识别模型。For example, the display device 200 may identify the head area in the image associated with the user by detecting the target shape in the image data associated with the user and the screen layout characteristics in the shape. From the head area, use the expression recognition model to recognize the current user's expression. Among them, the facial expression recognition model can be obtained by training an artificial intelligence model with sample data. That is, by inputting a large amount of sample data with expression labels into the initial model, and setting the model output result as the classification probability of the image belonging to a specific expression category, and then calculating the error between the classification probability and the expression label, and adjusting according to the error backpropagation Model parameters, so that the output result of the model is gradually the same as the label result, and an expression recognition model is obtained.
在表情识别模型输出当前图像针对某一表情的分类概率后,可以将分类概率最高的表情作为用户的表情,如微笑。再按照识别出的用户的表情在预设数据库中匹配相同类型的表情模型,即标准微笑模型,并使用标准微笑模型替换渲染模型中的面部区域,使渲染模型能够展示微笑表情。After the expression recognition model outputs the classification probability of the current image for a certain expression, the expression with the highest classification probability can be used as the user's expression, such as a smile. Then match the same type of expression model in the preset database according to the recognized user's expression, that is, the standard smile model, and use the standard smile model to replace the facial area in the rendering model, so that the rendering model can display smiling expressions.
可见,上述实施例可以通过匹配标准表情模型的方式,快速替换渲染模型中的面部区域,使显示设备200无需大量修改模型面部参数,减少数据处理量。并且可以获得较好的面部表情追踪时效,提高显示效果。It can be seen that the above embodiment can quickly replace the facial area in the rendering model by matching the standard expression model, so that the display device 200 does not need to modify the facial parameters of the model to reduce the amount of data processing. In addition, better facial expression tracking timeliness can be obtained, and the display effect can be improved.
基于上述虚拟试衣方法,在本公开的部分实施例中还提供一种显示设备200包括:显示器260、摄像头、通信器220以及控制器250。其中,显示器260被配置为显示图像和/或用户界面;摄像头被配置为实时采集图像数据,图像数据包括与用户关联的图像;通信器220被配置为与服务器400建立通信连接,服务器400内置模型重建应用,用于根据与用户关联的图像数据生成人体模型;控制器250被配置为执行以下程序步骤:Based on the above virtual fitting method, some embodiments of the present disclosure further provide a display device 200 including: a display 260 , a camera, a communicator 220 and a controller 250 . Wherein, the display 260 is configured to display images and/or user interfaces; the camera is configured to collect image data in real time, and the image data includes images associated with the user; the communicator 220 is configured to establish a communication connection with the server 400, and the server 400 has a built-in model a reconstruction application for generating a mannequin from image data associated with a user; the controller 250 is configured to perform the following procedural steps:
获取与用户关联的图像数据;Get the image data associated with the user;
将与用户关联的图像数据发送给服务器,以使得服务器根据与用户关联的图像数据生成并发送人体模型;sending the image data associated with the user to the server, so that the server generates and sends the mannequin according to the image data associated with the user;
为人体模型添加衣物素材,以合成渲染模型;Add clothing material to the mannequin to synthesize the rendered model;
从与用户关联的图像数据中提取动作参数,以及按照动作参数调节渲染模型的模型姿态,以渲染出试衣画面。Extract action parameters from the image data associated with the user, and adjust the model pose of the rendering model according to the action parameters to render a fitting picture.
如图13所示,图13示出本公开一些实施例的示例性虚拟试衣系统数据交互时序图。在一些实施例中,本公开还提供一种虚拟试衣系统,包括:显示设备200、图像采集装置500以及服务器400;其中,图像采集装置500连接显示设备200,显示设备200与服务器400建立通信连接。As shown in FIG. 13 , FIG. 13 shows an exemplary virtual fitting system data interaction sequence diagram of some embodiments of the present disclosure. In some embodiments, the present disclosure also provides a virtual fitting system, including: a display device 200, an image acquisition device 500, and a server 400; wherein, the image acquisition device 500 is connected to the display device 200, and the display device 200 establishes communication with the server 400 connect.
S1300、图像采集装置500接收来自用户的与用户关联的图像。S1300. The image acquisition device 500 receives an image associated with the user from the user.
S1301、图像采集装置500对与用户关联的图像执行图像信号处理,以生成与用户关联的图像数据。S1301. The image acquisition apparatus 500 performs image signal processing on an image associated with a user, so as to generate image data associated with the user.
S1302、图像采集装置500将与用户关联的图像数据发送至显示设备200。S1302. The image acquisition apparatus 500 sends the image data associated with the user to the display device 200.
S1303、显示设备200将与用户关联的图像数据发送至服务器400。S1303. The display device 200 sends the image data associated with the user to the server 400.
S1304、服务器400运行模型重建应用;根据与用户关联的图像数据生成人体模型。S1304. The server 400 runs a model reconstruction application; generates a human body model according to the image data associated with the user.
S1305、服务器400向显示设备200发送人体模型。S1305. The server 400 sends the human body model to the display device 200 .
S1306、显示设备200向人体模型添加衣物素材,以合成渲染模型;以及,从与用户关联的图像数据中提取动作参数,并按照动作参数调节渲染模型的模型姿态,以渲染出试衣画面。S1306. The display device 200 adds clothing materials to the human body model to synthesize a rendered model; and extracts action parameters from image data associated with the user, and adjusts the model pose of the rendered model according to the action parameters to render a fitting picture.
S1307、显示设备200向用户关联的设备发送试衣画面。S1307. The display device 200 sends a fitting screen to the device associated with the user.
可见,在上述实施例中,整个虚拟试衣系统可以包括:云端服务器400、本地显示设备200以及图像采集装置500。其中,云端服务器400可以负责人体重建,包括头部重建、身体重建、表情识别、头发试戴等基础算法模块,为显示设备200提供人体模型、表情驱动参数、试戴头发等功能支持。服务器400可以依赖图像采集装置500采集的数据作为输入,并将建模处理结果输出给显示设备200支持应用。It can be seen that, in the above embodiments, the entire virtual fitting system may include: a cloud server 400 , a local display device 200 and an image acquisition device 500 . Among them, the cloud server 400 can be responsible for human body reconstruction, including basic algorithm modules such as head reconstruction, body reconstruction, expression recognition, hair try-on, etc., and provides functional support for the display device 200 such as human body model, expression driving parameters, and hair try-on. The server 400 may rely on the data collected by the image collection device 500 as input, and output the modeling processing results to the display device 200 to support applications.
图像采集装置500则负责提供采集数据,包括图像信号处理(Image Signal Processing,ISP)调试、RGBD流、3D人体骨架关键点、人头部检测、多目标跟踪数据等,从而通过 OpenNI进行通信传输,为显示设备200提供RGBD图像数据、肢体驱动、人体跟踪、人脸重建素材、表情驱动素材的功能支持。显示设备200端则可以负责人体模型的渲染、衣服素材显示及渲染、动作驱动以及本地参数调整等功能。The image acquisition device 500 is responsible for providing acquisition data, including image signal processing (Image Signal Processing, ISP) debugging, RGBD stream, key points of 3D human skeleton, human head detection, multi-target tracking data, etc., so as to communicate and transmit through OpenNI, Provide the display device 200 with functional support for RGBD image data, limb driving, human body tracking, face reconstruction material, and expression driving material. The display device 200 can be responsible for functions such as rendering of the human body model, display and rendering of clothing materials, motion driving, and local parameter adjustment.
由此可知,本公开提供的虚拟试衣系统可以在使用中实时采集与用户关联的图像数据,并将与用户关联的图像数据发送给服务器400生成人体模型。再向人体模型添加衣物素材合成渲染模型,以及实时提取人物动作,从而按照动作参数调节渲染模型的模型姿态,形成试衣画面。虚拟试衣系统可以实现动态3D虚拟试衣功能,并通过渲染模型实时展示用户动作,达到衣随人动、穿搭推荐的使用效果,解决传统虚拟试衣方法无法实时展示试衣效果的问题。It can be seen that the virtual fitting system provided by the present disclosure can collect image data associated with the user in real time during use, and send the image data associated with the user to the server 400 to generate a human body model. Then add clothing materials to the human body model to synthesize the rendering model, and extract the character's actions in real time, so as to adjust the model pose of the rendering model according to the action parameters to form a fitting screen. The virtual fitting system can realize the dynamic 3D virtual fitting function, and display the user's actions in real time through the rendering model, so as to achieve the effect of clothes moving with people and recommended outfits, and solve the problem that the traditional virtual fitting method cannot display the fitting effect in real time.
本公开提供的实施例之间的相似部分相互参见即可,以上提供的具体实施方式只是本公开总的构思下的几个示例,并不构成本公开保护范围的限定。对于本领域的技术人员而言,在不付出创造性劳动的前提下依据本公开方案所扩展出的任何其他实施方式都属于本公开的保护范围。Similar parts between the embodiments provided in the present disclosure can be referred to each other, and the specific implementation manners provided above are only a few examples under the general concept of the present disclosure, and do not constitute a limitation of the protection scope of the present disclosure. For those skilled in the art, any other implementations extended according to the solutions of the present disclosure on the premise of not paying creative efforts belong to the protection scope of the present disclosure.

Claims (10)

  1. 一种显示设备,包括:A display device comprising:
    显示器,被配置为显示图像和/或用户界面;a display configured to display images and/or a user interface;
    摄像头,被配置为实时采集图像数据,所述图像数据包括与用户关联的图像;a camera configured to collect image data in real time, the image data including images associated with the user;
    通信器,被配置为与服务器建立通信连接,所述服务器内置模型重建应用,用于根据与用户关联的图像数据生成人体模型;The communicator is configured to establish a communication connection with a server, and the server has a built-in model reconstruction application for generating a human body model according to the image data associated with the user;
    控制器,与所述显示器、摄像头和通信器连接,被配置为:a controller, coupled to said display, camera and communicator, configured to:
    获取所述与用户关联的图像数据;Obtain the image data associated with the user;
    将所述与用户关联的图像数据发送给所述服务器,以使得所述服务器根据所述与用户关联的图像数据生成人体模型并向所述控制器发送所述人体模型;sending the image data associated with the user to the server, so that the server generates a mannequin according to the image data associated with the user and sends the mannequin to the controller;
    为所述人体模型添加衣物素材,以合成渲染模型;Adding clothing materials to the human body model to synthesize the rendered model;
    从所述与用户关联的图像数据中提取动作参数,以及按照所述动作参数调节所述渲染模型的模型姿态,以渲染出试衣画面。Extracting action parameters from the image data associated with the user, and adjusting the model pose of the rendering model according to the action parameters, so as to render a fitting picture.
  2. 根据权利要求1所述的显示设备,所述控制器被进一步配置为:The display device according to claim 1, the controller is further configured to:
    从所述与用户关联的图像数据中提取初始关键帧图像;extracting an initial keyframe image from said image data associated with the user;
    在所述初始关键帧图像中提取图像深度参数;extracting image depth parameters in the initial key frame image;
    将所述图像深度参数发送给所述服务器,以使所述服务器根据所述图像深度参数设置人体模型参数。Sending the image depth parameter to the server, so that the server sets the human body model parameter according to the image depth parameter.
  3. 根据权利要求2所述的显示设备,所述控制器被进一步配置为:The display device according to claim 2, the controller is further configured to:
    从所述与用户关联的图像数据中提取实时关键帧图像;extracting real-time keyframe images from the image data associated with the user;
    对比所述初始关键帧图像和所述实时关键帧图像,以获得图像相似度;comparing the initial key frame image with the real-time key frame image to obtain image similarity;
    如果所述图像相似度大于或等于预设相似度阈值,从所述与用户关联的图像数据中提取动作参数;If the image similarity is greater than or equal to a preset similarity threshold, extracting action parameters from the image data associated with the user;
    如果所述图像相似度小于所述预设相似度阈值,从所述实时关键帧图像中提取图像深度参数;If the image similarity is less than the preset similarity threshold, extracting an image depth parameter from the real-time key frame image;
    将所述图像深度参数发送给所述服务器,以使得所述服务器根据所述图像深度参数重建人体模型。Sending the image depth parameters to the server, so that the server reconstructs a human body model according to the image depth parameters.
  4. 根据权利要求3所述的显示设备,所述控制器被进一步配置为:The display device according to claim 3, the controller is further configured to:
    在所述初始关键帧图像中提取图像深度参数的后,记录所述初始关键帧图像;After extracting the image depth parameters in the initial key frame image, recording the initial key frame image;
    在从所述与用户关联的图像数据中提取动作参数后,或者在将所述图像深度参数发送给所述服务器后,所述控制器进一步被配置为:After extracting the action parameters from the image data associated with the user, or after sending the image depth parameters to the server, the controller is further configured to:
    使用所述实时关键帧图像替换记录的所述初始关键帧图像。The recorded initial key frame image is replaced with the real-time key frame image.
  5. 根据权利要求1所述的显示设备,所述控制器被进一步配置为:The display device according to claim 1, the controller is further configured to:
    从所述与用户关联的图像数据中识别人像目标;identifying a portrait subject from said image data associated with the user;
    为所述人像目标添加骨骼关键点,以生成骨骼参数;Add skeletal key points for the portrait target to generate skeletal parameters;
    向所述服务器发送所述骨骼参数,以使所述服务器根据所述骨骼参数设置所述人体模型的关节点位置。Sending the skeleton parameters to the server, so that the server sets the joint point positions of the human body model according to the skeleton parameters.
  6. 根据权利要求1所述的显示设备,所述控制器被进一步配置为:The display device according to claim 1, the controller is further configured to:
    获取用户输入的用于选定衣物的选择指令,所述选择指令中指定至少一个目标衣物;Obtaining a selection instruction input by the user for selecting clothing, where at least one target clothing is specified in the selection instruction;
    响应于所述选择指令,从衣物素材库中提取目标衣物素材;Responding to the selection instruction, extracting the target clothing material from the clothing material library;
    按照预设穿搭推荐规则,根据所述目标衣物素材在所述衣物素材库中匹配关联衣物素材;Matching associated clothing materials in the clothing material library according to the target clothing materials according to the preset recommendation rules for wearing;
    向所述人体模型添加所述目标衣物素材和所述关联衣物素材。Adding the target clothing material and the associated clothing material to the mannequin.
  7. 根据权利要求1所述的显示设备,所述控制器被进一步配置为:The display device according to claim 1, the controller is further configured to:
    遍历所述与用户关联的图像数据中每帧图像的骨骼关键点;Traversing the skeletal key points of each frame of image in the image data associated with the user;
    对比相邻两帧图像中的骨骼关键点位置,以获得每个骨骼关键点的移动距离;Compare the position of the bone key points in two adjacent frames of images to obtain the movement distance of each bone key point;
    根据所述移动距离移动所述人体模型的关节点位置。Move the joint point positions of the human body model according to the moving distance.
  8. 根据权利要求1所述的显示设备,所述控制器被进一步配置为:The display device according to claim 1, the controller is further configured to:
    在所述与用户关联的图像数据中识别头部区域;identifying a head region in said image data associated with the user;
    在所述头部区域中检测用户的表情;detecting a user's expression in said head region;
    根据所述用户的表情所属的表情类型,在预设表情库中匹配与所述表情类型相同的表情模型;According to the expression type to which the user's expression belongs, match an expression model identical to the expression type in a preset expression library;
    使用所述表情模型替换所述渲染模型中的面部区域。The facial region in the rendered model is replaced with the expression model.
  9. 一种虚拟试衣系统,包括:显示设备、图像采集装置以及服务器;其中,所述图像采集装置连接所述显示设备,所述显示设备与所述服务器建立通信连接;A virtual fitting system, comprising: a display device, an image acquisition device, and a server; wherein the image acquisition device is connected to the display device, and the display device establishes a communication connection with the server;
    所述图像采集装置被配置为实时采集与用户关联的图像,以及对所述与用户关联的图像进行图像信号处理,以生成与用户关联的图像数据;所述图像采集装置还被配置为向所述显示设备发送所述与用户关联的图像数据;The image acquisition device is configured to collect images associated with the user in real time, and perform image signal processing on the image associated with the user to generate image data associated with the user; the image acquisition device is also configured to provide The display device sends the image data associated with the user;
    所述显示设备被配置为获取所述与用户关联的图像数据,以及将所述与用户关联的图像数据发送给所述服务器;The display device is configured to acquire the image data associated with the user, and send the image data associated with the user to the server;
    所述服务器中内置模型重建应用,所述服务器被配置为接收所述与用户关联的图像数据,以及运行所述模型重建应用;根据所述与用户关联的图像数据生成人体模型,所述服务器还被配置为向所述显示设备发送所述人体模型;A built-in model reconstruction application in the server, the server is configured to receive the image data associated with the user, and run the model reconstruction application; generate a human body model according to the image data associated with the user, and the server also configured to send the mannequin to the display device;
    所述显示设备还被配置为向所述人体模型添加衣物素材,以合成渲染模型;以及,从所述与用户关联的图像数据中提取动作参数,并按照所述动作参数调节所述渲染模型的模型姿态,以渲染出试衣画面。The display device is further configured to add clothing materials to the human body model to synthesize a rendered model; and extract motion parameters from the image data associated with the user, and adjust the rendering model according to the motion parameters. Model pose to render the fitting screen.
  10. 一种虚拟试衣方法,应用于虚拟试衣系统,所述虚拟试衣系统包括:显示设备、图像采集装置以及服务器;其中,所述图像采集装置连接所述显示设备,所述显示设备与所述服务器建立通信连接;所述虚拟试衣方法包括:A virtual fitting method, applied to a virtual fitting system, the virtual fitting system includes: a display device, an image acquisition device, and a server; wherein, the image acquisition device is connected to the display device, and the display device is connected to the display device Said server establishes communication connection; Described virtual fitting method comprises:
    所述图像采集装置实时采集与用户关联的图像,以及对所述与用户关联的图像进行图像信号处理,以生成与用户关联的图像数据;The image collection device collects images associated with the user in real time, and performs image signal processing on the image associated with the user to generate image data associated with the user;
    所述显示设备获取所述与用户关联的图像数据,以及将所述与用户关联的图像数据发送给所述服务器;The display device acquires the image data associated with the user, and sends the image data associated with the user to the server;
    所述服务器接收所述与用户关联的图像数据,根据所述与用户关联的图像数据生成人体模型,以及向所述显示设备发送所述人体模型;The server receives the image data associated with the user, generates a human body model according to the image data associated with the user, and sends the human body model to the display device;
    所述显示设备为所述人体模型添加衣物素材,以合成渲染模型;The display device adds clothing material to the human body model to synthesize a rendered model;
    所述显示设备从所述与用户关联的图像数据中提取动作参数,并按照所述动作参数调节所述渲染模型的模型姿态,以渲染出试衣画面。The display device extracts action parameters from the image data associated with the user, and adjusts the model pose of the rendering model according to the action parameters, so as to render a fitting picture.
PCT/CN2022/128392 2022-01-17 2022-10-28 Display device, and virtual fitting system and method WO2023134269A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210051018.9A CN116523579A (en) 2022-01-17 2022-01-17 Display equipment, virtual fitting system and method
CN202210051018.9 2022-01-17

Publications (1)

Publication Number Publication Date
WO2023134269A1 true WO2023134269A1 (en) 2023-07-20

Family

ID=87280075

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/128392 WO2023134269A1 (en) 2022-01-17 2022-10-28 Display device, and virtual fitting system and method

Country Status (2)

Country Link
CN (1) CN116523579A (en)
WO (1) WO2023134269A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117635883A (en) * 2023-11-28 2024-03-01 广州恒沙数字科技有限公司 Virtual fitting generation method and system based on human skeleton posture
CN117649283B (en) * 2023-12-14 2024-05-14 杭州抽象派数字科技有限公司 Virtual fitting system and virtual fitting method

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116880948A (en) * 2023-09-07 2023-10-13 深圳星坊科技有限公司 Jewelry virtual try-on display method, jewelry virtual try-on display device, computer equipment and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1074944A2 (en) * 1999-07-12 2001-02-07 Styleclick.Com Inc. Method and apparatus for combining and visualizing items of apparel
CN102156810A (en) * 2011-03-30 2011-08-17 北京触角科技有限公司 Augmented reality real-time virtual fitting system and method thereof
CN102298797A (en) * 2011-08-31 2011-12-28 深圳市美丽同盟科技有限公司 Three-dimensional virtual fitting method, device and system
CN105825407A (en) * 2016-03-31 2016-08-03 上海晋荣智能科技有限公司 Virtual fitting mirror system

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1074944A2 (en) * 1999-07-12 2001-02-07 Styleclick.Com Inc. Method and apparatus for combining and visualizing items of apparel
CN102156810A (en) * 2011-03-30 2011-08-17 北京触角科技有限公司 Augmented reality real-time virtual fitting system and method thereof
CN102298797A (en) * 2011-08-31 2011-12-28 深圳市美丽同盟科技有限公司 Three-dimensional virtual fitting method, device and system
CN105825407A (en) * 2016-03-31 2016-08-03 上海晋荣智能科技有限公司 Virtual fitting mirror system

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117635883A (en) * 2023-11-28 2024-03-01 广州恒沙数字科技有限公司 Virtual fitting generation method and system based on human skeleton posture
CN117649283B (en) * 2023-12-14 2024-05-14 杭州抽象派数字科技有限公司 Virtual fitting system and virtual fitting method

Also Published As

Publication number Publication date
CN116523579A (en) 2023-08-01

Similar Documents

Publication Publication Date Title
US20210177124A1 (en) Information processing apparatus, information processing method, and computer-readable storage medium
WO2023134269A1 (en) Display device, and virtual fitting system and method
US10109315B2 (en) Devices, systems and methods for auto-delay video presentation
US11908052B2 (en) System and method for digital makeup mirror
RU2668408C2 (en) Devices, systems and methods of virtualising mirror
US11798201B2 (en) Mirroring device with whole-body outfits
US9098873B2 (en) Motion-based interactive shopping environment
JP3984191B2 (en) Virtual makeup apparatus and method
JP4435809B2 (en) Virtual makeup apparatus and method
JP2019510297A (en) Virtual try-on to the user's true human body model
US20230281936A1 (en) Controlling interactive fashion based on body gestures
WO2016112346A1 (en) Devices, systems and methods for auto-delay video presentation
US11900506B2 (en) Controlling interactive fashion based on facial expressions
WO2014081394A1 (en) Method, apparatus and system for virtual clothes modelling
CN111199583B (en) Virtual content display method and device, terminal equipment and storage medium
US20240013463A1 (en) Applying animated 3d avatar in ar experiences
US20240096040A1 (en) Real-time upper-body garment exchange
Chen et al. 3D face reconstruction and gaze tracking in the HMD for virtual interaction
CN117292097B (en) AR try-on interactive experience method and system
Sénécal et al. Modelling life through time: cultural heritage case studies
Woodward et al. An interactive 3D video system for human facial reconstruction and expression modeling
US20240020901A1 (en) Method and application for animating computer generated images
US20240070969A1 (en) Multisensorial presentation of volumetric content
Shinkar et al. A Real Time Virtual Dressing Room Application using Opencv
Azam et al. A Survey on 3D Egocentric Human Pose Estimation

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22919903

Country of ref document: EP

Kind code of ref document: A1