WO2018036456A1 - Procédé et dispositif pour suivre et reconnaître une marchandise dans une image vidéo et afficher des informations de marchandise - Google Patents

Procédé et dispositif pour suivre et reconnaître une marchandise dans une image vidéo et afficher des informations de marchandise Download PDF

Info

Publication number
WO2018036456A1
WO2018036456A1 PCT/CN2017/098325 CN2017098325W WO2018036456A1 WO 2018036456 A1 WO2018036456 A1 WO 2018036456A1 CN 2017098325 W CN2017098325 W CN 2017098325W WO 2018036456 A1 WO2018036456 A1 WO 2018036456A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
panoramic video
video
virtualized
feature
Prior art date
Application number
PCT/CN2017/098325
Other languages
English (en)
Chinese (zh)
Inventor
郑浩
潘杰
张怡
丁航
Original Assignee
大辅科技(北京)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 大辅科技(北京)有限公司 filed Critical 大辅科技(北京)有限公司
Publication of WO2018036456A1 publication Critical patent/WO2018036456A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/858Linking data to content, e.g. by linking an URL to a video object, by creating a hotspot

Definitions

  • the present application relates to the field of image recognition, and in particular to a method and apparatus for tracking and identifying products in a video image and displaying product information.
  • video website refers to a website that allows Internet users to publish, browse and share video works online with the support of related technology platforms.
  • the well-known video websites include Youku, LeTV, iQiyi, etc.; usually video websites will also be launched.
  • Its own video client application also known as video client
  • video client is specially used to play video works provided by video websites on terminal devices such as mobile phones or personal computers, such as Youku video client and iQiyi video client.
  • the present invention provides a new mode of loading advertisements on video in order to improve the conversion rate of advertisements.
  • a method of tracking and identifying an item in a video image and displaying the item information comprising:
  • the region of the feature image in the live image is loaded with the product information as the product image region;
  • the product information is displayed.
  • the live image is virtualized to form a first virtualized image
  • the first virtualized image includes the live image a first virtual object corresponding to the medium feature image
  • the video is decomposed into a plurality of real-image images according to a frame, and in the plurality of first virtualized images formed by performing virtualization processing on the plurality of real-image images, the identifiers of the first virtual objects corresponding to the same object are consistent. .
  • the first virtualized image is located in a merchandise image area.
  • the second virtualized image is commodity information.
  • the second virtualized image is displayed during video playback. Since the video can be viewed as a combination of multi-frame live images, it can be seen that the second virtualized image follows the positional display of the feature image in the video.
  • the second virtualized image is displayed while the video is paused.
  • displaying product information during video playback can interfere with their viewing of the video.
  • the second virtualized image may be set to be transparent when the video is played, and displayed when paused.
  • an apparatus for tracking and identifying an item in a video image and displaying the item information characterized by comprising
  • An image recognition unit configured to acquire a real-life image and identify a feature image in the real-life image
  • a display unit for displaying video and product information
  • a feature database for storing product feature identifiers and corresponding product information
  • the commodity information loading unit is configured to load the region of the feature image in the live image as the product image region with the corresponding product information.
  • the product information loading unit virtualizes the real-life image acquired by the image recognition unit to form a product image region.
  • the first virtualized image includes a first virtual object corresponding to the feature image in the live view image; an identifier is added to the preset position of the first virtual object in the first virtualized image, and the generated a second virtualized image of the item information, wherein the identification is for identifying the first virtual object.
  • the image recognition unit decomposes the video into a plurality of real-life images according to the frame, and in the plurality of first virtualized images formed by the virtualizing the plurality of real-image images, the first virtual object corresponding to the same object
  • the logo is consistent.
  • the item information includes one or more of a URL, a product introduction, and a brand introduction.
  • a client of a panoramic video system includes: a data receiving module and a data display module, wherein: the data receiving module is configured to receive video data and send the data to the data display module, where The video data includes a panoramic video and a corresponding VR image or a specific image, wherein the VR image is accompanied by a hyperlink; the data display module is configured to display the panoramic video and the corresponding VR image or a specific image thereof, and A request signal generated by the user clicking the hyperlink is received and transmitted to the outside.
  • the panoramic video is a spherical panoramic video or a cubic panoramic video, the specific image being displayed at the top or bottom of the spherical panoramic video or the cube panoramic video.
  • the panoramic video is a cylindrical panoramic video, the particular image being displayed above or below the cylindrical panoramic video.
  • the data display module is further configured to jump to display the specific image from the panoramic video.
  • a server for a panoramic video system includes: a VR generating module, a request processing module, and a data sending module, wherein the VR generating module is configured to generate and send a corresponding VR image based on the panoramic video.
  • a data sending module configured to receive a request signal from the outside and perform a corresponding operation on the request signal and generate a specific image of the corresponding operation, and send the specific image to the data a sending module; the data sending module, configured to send video data to the outside, wherein the video data includes the panoramic video and the corresponding VR image or the specific image.
  • a panoramic video system comprising a client as described above and a server as described above.
  • it is a panoramic video map system, the panoramic video being a street panoramic video.
  • the street panoramic video comprises a street real scene and a street map.
  • the VR image is a VR image of a commodity.
  • the specific image is an advertisement of the specific item or a website of the specific item.
  • a panoramic video product introduction system including:
  • An image recognition unit configured to acquire a real-life image and identify a feature image in the real-life image
  • a display unit for displaying video and product information
  • a feature database for storing product feature identifiers and corresponding product information
  • a virtual image generating unit that generates a virtual image based on the feature image and superimposes on the real image
  • the commodity information loading unit is configured to load the region of the feature image in the live image as the product image region with the corresponding product information.
  • FIG. 1 is a flow diagram of a method of tracking an item in a video image and displaying item information, in accordance with some embodiments of the present invention.
  • FIG. 2 is a schematic diagram of an apparatus for tracking an item in a video image and displaying item information, in accordance with some embodiments of the present invention.
  • FIG. 3 is a panoramic video system in accordance with an embodiment of the present invention.
  • FIG. 4 is a client of a panoramic video system in accordance with a particular embodiment of the present invention.
  • Figure 5 is a server of a panoramic video system in accordance with an embodiment of the present invention.
  • FIG. 1 is a flow diagram of a method of tracking an item in a video image and displaying item information, in accordance with some embodiments of the present invention.
  • step S110 acquiring a real-life image.
  • the virtual reality device or the augmented reality device can capture the current scene through its own image acquisition unit to generate a real-life image.
  • the virtual reality device can also receive real-life images sent by other electronic devices.
  • Step S120 Identify the feature image in the live view image.
  • the video is broken down into frames into a continuous still image. Calculate the image for each frame of still image
  • the sharpness of the static image is obtained by the edge point and/or the manner in which the image sharpness is calculated.
  • the image is divided into a plurality of image regions by using an image segmentation technique; for each image region, a feature value for describing the property of the image region is obtained, and it is determined whether the feature value meets a preset feature of the product image region. If yes, it is determined that the image area is a product image area, and the position of the area in the still image is the position of the product contained therein in the still image.
  • Step S130 Determine whether the feature image matches the product feature in the database.
  • the image features in the product image area are extracted as reference features; the candidate product features matching the reference features are searched in a previously generated feature database, and the product information corresponding to the candidate product features that are successfully matched is obtained.
  • Step S140 When the feature image matches the product feature in the database, the region of the feature image in the live image is loaded with the product information as the product image region.
  • the acquired real-life image is virtualized to form a first virtualized image.
  • the first virtualized image includes a virtual object corresponding to the feature image in the live image.
  • the virtual reality device virtualizes the acquired real-life image to form a first virtualized image, wherein the first virtualized image includes a virtual object corresponding to all or part of the objects in the real-life image.
  • a virtual object can present a three-dimensional state, or a planar state.
  • the virtual object in the first virtualized image formed by the virtual reality device is a simulation of the corresponding object in the real image.
  • the augmented reality device performs virtualization processing on the acquired real-life image, specifically, adding a corresponding virtual object to all or part of the feature image in the real-life image, and the added virtual object constitutes the first virtualized image.
  • An identifier is added to the preset position of the first virtual object in the first virtualized image to form a second virtualized image. Wherein the identifier is used to identify the first virtual object.
  • the first virtual object is a virtual object corresponding to a specific image in the live image, and the first virtual object may be one or more.
  • the first virtual object may be a virtual object corresponding to an object in the real image, and may be a virtual object corresponding to a certain type of object in the real image, or may be a virtual corresponding to one or more objects selected by the user.
  • Object may be a virtual object corresponding to an object in the real image, and may be a virtual object corresponding to a certain type of object in the real image, or may be a virtual corresponding to one or more objects selected by the user.
  • the identifier is used to identify the first virtual object, and the user is convenient to distinguish the first virtual object from the other virtual object, so that the user can track the first virtual object.
  • Step S150 Display the item information (second virtualized image).
  • the display unit of the virtual reality device usually displays a closed display effect, that is, the user can hardly see the real environment through the display unit of the virtual reality device, and often only sees the second virtualization displayed by the display unit of the virtual reality device. image.
  • the display unit of the augmented reality device presents a non-closed display effect, that is, the user can view the real environment through the display unit of the augmented reality device while viewing the image displayed by the augmented reality device through the display unit, and the real scene is
  • the second virtualized image is superimposed and displayed, which can produce a visual effect combining virtual and real.
  • the electronic device performs virtualization processing on the acquired real-life image to form a first virtualized image, and adds a logo to the preset position of the first virtual object to form a second virtualized image in the first virtualized image, and
  • the second virtualized image is displayed by the display unit.
  • adding the identifier to the preset position of the first virtual object in the first virtualized image may be: adding a preset icon to the preset position of the first virtual object in the first virtualized image.
  • the first virtual object and other virtual objects can be visually distinguished by the icon added at the preset position of the first virtual object.
  • the position of the object in the live image changes, the position of the virtual object corresponding to the object is also changed correspondingly in the first virtualized image formed based on the real image, and the icon at the preset position of the virtual object is Visual objects can be visually tracked.
  • the method shown in FIG. 1 may be improved. Specifically, in the plurality of first virtualized images formed by performing virtualization processing on the plurality of real-life images, the identifiers of the first virtual objects corresponding to the same object are consistent.
  • the virtual reality device or the augmented reality device generally acquires the real-life image according to the preset time interval, and virtualizes the acquired real-life image to form a first virtualized image, and displays the first virtualized image through the display unit of the device, thereby The user presents a coherent image.
  • the identifiers added to the first virtual objects corresponding to the same object are consistent. That is to say, in the plurality of second virtualized images formed for the same scene, the identifiers of the virtual objects corresponding to the same object remain unchanged.
  • the identifiers of the virtual objects corresponding to different objects may be further set to be different.
  • the user can not only distinguish each first virtual object, but also intuitively track the first virtual object according to the identifier of each first virtual object.
  • FIG. 2 is a schematic diagram of an apparatus for tracking an item in a video image and displaying item information, in accordance with some embodiments of the present invention.
  • the image recognition unit 101 is configured to acquire a real-life image and identify a feature image in the real-life image.
  • the real-life image may be captured by the image acquisition unit of the current scene to generate a real-life image, or may be a captured video image.
  • a video or live image
  • the sharpness of the still image is obtained by calculating the image edge points and/or calculating the image sharpness.
  • the image is divided into a plurality of image regions by using an image segmentation technique; for each image region, Obtaining a feature value for describing a property of the image region, determining whether the feature value meets a feature of a preset product image region; if yes, determining that the image region is a product image region, and the region is in the static image
  • the location is the location of the item contained therein in the still image.
  • the feature database 103 is configured to store the product feature identifier and the corresponding product information.
  • the image recognition unit 101 extracts an image feature in the product image region as a reference feature; searches for a candidate product feature matching the reference feature in the feature database 103 generated in advance, and acquires product information corresponding to the candidate product feature that is successfully matched. .
  • the product information loading unit 104 is configured to load the region of the feature image in the live image as the product image region with the corresponding product information.
  • the product information loading unit 104 performs virtualization processing on the real image acquired by the image recognition unit, and forms a first virtualized image in the product image region.
  • a virtualized image includes a first virtual object corresponding to the feature image in the live view image; an identifier is added to the preset position of the first virtual object in the first virtualized image, and a second virtualized image including the product information is generated Wherein the identifier is for identifying the first virtual object.
  • the item information includes one or more of a URL, a product introduction, and a brand introduction.
  • the image recognition unit 101 decomposes the video into a plurality of real-life images according to a frame.
  • the identifier of the first virtual object corresponding to the same object remains. Consistent.
  • the display unit 102 is configured to display video and product information.
  • a display unit of a virtual reality device (such as a VR helmet, VR glasses, or a VR box) usually displays a closed display effect, which means that it is difficult for a user to see the real environment through the display unit of the virtual reality device, and often only see the virtual environment.
  • the second virtualized image displayed by the display unit of the real device is usually displayed by the display unit of the real device.
  • the display unit of the augmented reality device (such as a smart phone, a Pad, an AR helmet, etc.) presents a non-closed display effect, that is, the user can view the real environment through the display unit of the augmented reality device while viewing the augmented reality device through the display unit.
  • the displayed image is superimposed and displayed for the user on the real scene and the second virtualized image, and can produce a visual effect combining virtual and real.
  • the panoramic video system is composed of panoramic video editing and browsing software and panoramic video capture and transmission equipment. It can provide people with rich panoramic images and combine modern network technology. It is an effective fusion of video capture and virtual reality technology. It can be easily used, and the collection device is relatively economical to use, and finally generates a high-information digital media solution, which can provide real-time for people. Video information.
  • VR Virtual Reality
  • FIG. 3 is a panoramic video system in accordance with an embodiment of the present invention.
  • a panoramic video system 100 is provided in accordance with an embodiment of the present invention, including a client 110 and a server 120.
  • a client 110 of a panoramic video system 100 is provided in accordance with an embodiment of the present invention.
  • the client 110 of the panoramic video system includes a data receiving module 111 and a data display module 112.
  • the data receiving module 111 is configured to receive video data and send the data to the data display module, where the video data includes a panoramic video and a corresponding VR image or a specific image, wherein the VR image is accompanied by a hyperlink; the data display module 112.
  • the data receiving module 111 can be implemented in various manners. Specifically, for example, the data receiving module 111 includes three sub-modules, namely, a network receiving feedback sub-module 114, a decoding video synchronization sub-module 115, and a real-time stitching sub-module 116.
  • the network receiving feedback module is responsible for receiving video data transmitted from the server
  • the decoding video synchronization module is responsible for converting the video data into multiple segments of video in the form of panoramic video, and synchronizing the video segments
  • the real-time stitching module is to segment each segment.
  • the video is stitched in real time under the guidance of certain parameters, and finally a complete panoramic video is synthesized.
  • the data receiving module 111 directly receives the panoramic video from the outside, that is, directly transmits the panoramic video to the data display module 112 without performing any data processing.
  • the data display module 112 provides the user with a three-dimensional viewing interface to present the panoramic video to the people.
  • the data display module also displays the VR image generated according to the panoramic video.
  • the user can also see the VR image, and the VR image is accompanied by a hyperlink, that is, the user can click on the VR image to enable the panoramic video system to execute and super
  • Corresponding specific operations are linked, and in this particular embodiment, for example, the data display module displays a particular image.
  • the data display module is further configured to receive a request signal generated by the user clicking the VR image, and send the request signal to the outside.
  • the user can pass the point Clicking on the VR image for further content or interacting with the client in other forms enhances the richness and interactivity of the panoramic video.
  • cylindrical panoramic video There are three main display modes for panoramic video, cylindrical panoramic video, spherical panoramic video and cube panoramic video.
  • cylindrical panoramic video best meets the needs of VR images and roaming systems, and is relatively easy to create and convenient to use.
  • the design method of the panoramic photo is mainly used.
  • the particular image is displayed above or below the cylindrical panoramic video.
  • the particular image does not affect the user's main view of the panoramic video.
  • the panoramic video is a spherical panoramic video or a cube panoramic video, the particular image being displayed at the top or bottom of the spherical panoramic video or cube panoramic video.
  • the particular image does not affect the user's primary view of the panoramic video.
  • the data display module is further configured to jump from the panoramic video to display the particular image.
  • the panoramic video jumps to a specific image, the user can view the specific image more conveniently, and it is also easier to present a specific image more abundantly.
  • FIG. 5 is a server of a panoramic video system in accordance with an embodiment of the present invention.
  • a server 120 of a panoramic video system 100 is provided in accordance with an embodiment of the present invention.
  • a server 120 of a panoramic video system includes a VR generating module 121, a request processing module 122, and a data sending module 123.
  • the VR generating module 121 is configured to generate a corresponding VR image based on the panoramic video and send it to the data sending module 123.
  • the request processing module 122 is configured to receive a request signal from the outside and perform a corresponding operation on the request signal and generate a specific image of the corresponding operation, and the specific image is sent to the data sending module 123; the data sending module 123 is configured to send video data to the outside, wherein the video data includes the panoramic video and the The corresponding VR image or the specific image is described.
  • the server 120 of the panoramic video system receives the request signal from the external data, such as the data display module 112 of the client 110 of the panoramic video system, by requesting the processing module 122 and executes the request signal for the request signal.
  • the external data such as the data display module 112 of the client 110 of the panoramic video system
  • the server 120 of the panoramic video system is based on the panoramic video by the VR generating module 121 Generate their corresponding
  • the VR image is sent to the data sending module 123; the data sending module 123 sends video data to the external data receiving module 111 of the client 110, for example, the panoramic video system, wherein the video data includes the panoramic video and the Its corresponding VR image or the particular image.
  • the data sending module 123 can be implemented in various manners. Specifically, for example, the data sending module 123 includes three sub-modules, namely, a network receiving feedback sub-module, a decoding video synchronization sub-module, and a real-time stitching sub-module.
  • the network receiving feedback module is responsible for receiving video data transmitted from the outside
  • the decoding video synchronization module is responsible for converting the video data into multiple pieces of video in the form of panoramic video, and synchronizing each piece of video
  • the real-time stitching module is to segment each segment.
  • the video is stitched in real time under the guidance of certain parameters, and finally a complete panoramic video is synthesized.
  • the data transmitting module 123 directly transmits the video data to the outside, that is, directly transmits the video data to the data receiving module 111 without performing any data processing.
  • FIG. 3 is a panoramic video system according to a specific embodiment of the present invention.
  • a panoramic video system 100 including a client 110 and a server 120 is provided according to an embodiment of the present invention. .
  • a panoramic video system in accordance with a particular embodiment of the present invention is a panoramic video map system such that the panoramic video is a street panoramic video.
  • the street panoramic video includes a street real scene and a street map.
  • the existing electronic map only includes panoramic photos, and the invention is combined with an electronic map to provide a panoramic video for the user, thereby providing more accurate information.
  • the VR image is a VR image of a commodity.
  • the VR image of the product may be a VR image of the product or a VR image of the merchant. in particular,
  • the specific image in the client 110 or the server 120 of a panoramic video system is an advertisement of the specific item or a website of the specific item.
  • a VR-based panoramic video map or an AR-based panoramic video map can be developed.
  • a panoramic video in a predetermined area is pre-photographed.
  • the bakery matches the products sold in the bakery;
  • the sports store matches the products sold in the store;
  • the museum matches the information in the museum.
  • the first virtualized image is set as needed, for example, a product image is displayed. Considering the user experience, the content of the first virtualized image is as concise as possible. It is also possible to select a push image based on the result of the big data calculation.
  • the first virtualized image is presented.
  • the second virtualized image is played, and the second virtualized image is the product information.
  • the scene in the predetermined area is set as a feature image to match the product features in the database.
  • the bakery matches the products sold in the bakery; the sports store matches the products sold in the store; the museum matches the information in the museum.
  • the first virtualized image is set as needed, for example, a product image is displayed. Considering the user experience, the content of the first virtualized image is as concise as possible. It is also possible to select a push image based on the result of the big data calculation.
  • the scene set as the feature image is recognized in real time, and the first virtualized image is presented after the scene appears in the line of sight for more than a preset time (for example, 2 seconds).
  • the second virtualized image is played, and the second virtualized image is the commodity information.
  • the embodiment of the present invention is not limited to a panoramic video map, and may be any video product or a system using video.
  • the technical solution described in the present invention may be adopted.
  • the presentation content of the first virtualized image and the second virtualized image may also be designed according to customer needs, and is not limited to being exemplified in the embodiments of the present specification.
  • the descriptions of the various embodiments are different, and the details that are not detailed in a certain embodiment can be referred to the related descriptions of other embodiments.
  • the disclosed apparatus may be implemented in other ways.
  • the device embodiments described above are merely illustrative.
  • the division of the unit is only a logical function division.
  • there may be another division manner for example, multiple units or components may be combined or may be Integrate into another system, or some features can be ignored or not executed.
  • the mutual coupling or direct coupling or communication connection shown or discussed may be an indirect coupling or communication connection through some interface, device or unit, and may be electrical or otherwise.
  • the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of the embodiment.
  • each functional unit in each embodiment of the present invention may be integrated into one processing unit, or each unit may exist physically separately, or two or more units may be integrated into one unit.
  • the above integrated unit can be implemented in the form of hardware or in the form of a software functional unit.
  • the above method according to the present invention can be implemented in hardware, firmware, or as software or computer code that can be stored in a recording medium such as a CD ROM, a RAM, a floppy disk, a hard disk, or a magneto-optical disk, or can be downloaded through a network.
  • a recording medium such as a CD ROM, a RAM, a floppy disk, a hard disk, or a magneto-optical disk, or can be downloaded through a network.
  • the computer code originally stored in a remote recording medium or non-transitory machine readable medium and to be stored in a local recording medium, whereby the methods described herein can be stored using a general purpose computer, a dedicated processor, or programmable or dedicated Such software processing on a recording medium of hardware such as an ASIC or an FPGA.
  • a computer, processor, microprocessor controller or programmable hardware includes storage components (eg, RAM, ROM, flash memory, etc.) that can store or receive software or computer code, when the software or computer code is The processing methods described herein are implemented when the processor or hardware is accessed and executed. Moreover, when a general purpose computer accesses code for implementing the processing shown herein, the execution of the code converts the general purpose computer into a special purpose computer for performing the processing shown herein.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Processing Or Creating Images (AREA)

Abstract

La présente invention concerne un procédé de suivi et de reconnaissance d'une marchandise dans une image vidéo et d'affichage d'informations de marchandise. Le procédé consiste à : obtenir des images réelles; reconnaître une image caractéristique dans les images réelles; lorsque l'image caractéristique correspond à une caractéristique de marchandise dans une base de données, utiliser une zone contenant l'image caractéristique dans les images réelles en tant que zone d'image de marchandise pour charger des informations de marchandise; et afficher les informations de marchandise. La présente invention concerne également un dispositif de suivi et de reconnaissance d'une marchandise dans une image vidéo et d'affichage d'informations de marchandise.
PCT/CN2017/098325 2016-08-22 2017-08-21 Procédé et dispositif pour suivre et reconnaître une marchandise dans une image vidéo et afficher des informations de marchandise WO2018036456A1 (fr)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
CN201610700943 2016-08-22
CN201610700962 2016-08-22
CN201610700943.4 2016-08-22
CN201610700962.7 2016-08-22

Publications (1)

Publication Number Publication Date
WO2018036456A1 true WO2018036456A1 (fr) 2018-03-01

Family

ID=61100084

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/098325 WO2018036456A1 (fr) 2016-08-22 2017-08-21 Procédé et dispositif pour suivre et reconnaître une marchandise dans une image vidéo et afficher des informations de marchandise

Country Status (2)

Country Link
CN (1) CN107633441A (fr)
WO (1) WO2018036456A1 (fr)

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108364209A (zh) * 2018-02-01 2018-08-03 北京京东金融科技控股有限公司 商品信息的展示方法、装置、介质及电子设备
CN110456901A (zh) * 2019-08-16 2019-11-15 上海电气集团股份有限公司 展会中对象展示的控制方法、系统、电子设备和存储介质
CN110648142A (zh) * 2018-06-07 2020-01-03 阿里巴巴集团控股有限公司 商品溯源链路信息处理方法、装置以及电子设备
CN110858134A (zh) * 2018-08-22 2020-03-03 阿里巴巴集团控股有限公司 数据、显示处理方法、装置、电子设备和存储介质
CN111353839A (zh) * 2018-12-21 2020-06-30 阿里巴巴集团控股有限公司 商品信息处理方法、直播商品的方法、装置及电子设备
CN111597863A (zh) * 2019-02-21 2020-08-28 顺丰科技有限公司 一种装卸率确定方法、系统、设备及存储介质
CN111865771A (zh) * 2018-08-08 2020-10-30 创新先进技术有限公司 消息发送方法及装置和电子设备
CN111935488A (zh) * 2019-05-13 2020-11-13 阿里巴巴集团控股有限公司 数据处理方法、信息显示方法、装置、服务器及终端设备
CN112132644A (zh) * 2020-08-21 2020-12-25 苏州合浩网络科技有限公司 一种vr商城的商品智能陈列方法及更新系统
US10970519B2 (en) 2019-04-16 2021-04-06 At&T Intellectual Property I, L.P. Validating objects in volumetric video presentations
US11012675B2 (en) 2019-04-16 2021-05-18 At&T Intellectual Property I, L.P. Automatic selection of viewpoint characteristics and trajectories in volumetric video presentations
CN112991553A (zh) * 2021-03-11 2021-06-18 深圳市慧鲤科技有限公司 信息展示方法及装置、电子设备和存储介质
US11074697B2 (en) 2019-04-16 2021-07-27 At&T Intellectual Property I, L.P. Selecting viewpoints for rendering in volumetric video presentations
US11153492B2 (en) 2019-04-16 2021-10-19 At&T Intellectual Property I, L.P. Selecting spectator viewpoints in volumetric video presentations of live events

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108198044B (zh) * 2018-01-30 2021-01-26 京东数字科技控股有限公司 商品信息的展示方法、装置、介质及电子设备
CN110136265A (zh) * 2018-02-02 2019-08-16 北京京东尚科信息技术有限公司 商品展示方法、装置、终端设备及零售系统
CN108920707B (zh) * 2018-07-20 2022-03-15 百度在线网络技术(北京)有限公司 用于标注信息的方法及装置
CN110858375B (zh) * 2018-08-22 2023-05-02 阿里巴巴集团控股有限公司 数据、显示处理方法、装置、电子设备和存储介质
CN109104632A (zh) * 2018-09-27 2018-12-28 聚好看科技股份有限公司 一种电视终端ar场景的实现方法及系统
CN109462730A (zh) * 2018-10-25 2019-03-12 百度在线网络技术(北京)有限公司 基于视频获取全景图的方法和装置
CN110881134B (zh) * 2019-11-01 2020-12-11 北京达佳互联信息技术有限公司 数据处理方法、装置、电子设备及存储介质
CN114051089B (zh) * 2021-10-12 2023-09-15 聚好看科技股份有限公司 一种全景视频中投放资源的方法及显示设备
CN114296548B (zh) * 2021-12-14 2023-03-24 杭州朱道实业有限公司 一种展览用智能移动识别信息系统

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103577788A (zh) * 2012-07-19 2014-02-12 华为终端有限公司 增强现实的实现方法和装置
CN103886027A (zh) * 2014-02-26 2014-06-25 四川长虹电器股份有限公司 扫描可视区域获取物品信息的方法和电视
CN105373938A (zh) * 2014-08-27 2016-03-02 阿里巴巴集团控股有限公司 识别视频图像中的商品和展示其信息的方法、装置及系统

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102063436A (zh) * 2009-11-18 2011-05-18 腾讯科技(深圳)有限公司 一种利用终端获取图像实现商品信息搜索的系统及方法
KR101995958B1 (ko) * 2012-11-28 2019-07-03 한국전자통신연구원 스마트 글래스 기반 영상 처리 장치 및 방법
CN105812680A (zh) * 2016-03-31 2016-07-27 联想(北京)有限公司 图像处理方法及电子设备

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103577788A (zh) * 2012-07-19 2014-02-12 华为终端有限公司 增强现实的实现方法和装置
CN103886027A (zh) * 2014-02-26 2014-06-25 四川长虹电器股份有限公司 扫描可视区域获取物品信息的方法和电视
CN105373938A (zh) * 2014-08-27 2016-03-02 阿里巴巴集团控股有限公司 识别视频图像中的商品和展示其信息的方法、装置及系统

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108364209A (zh) * 2018-02-01 2018-08-03 北京京东金融科技控股有限公司 商品信息的展示方法、装置、介质及电子设备
CN110648142A (zh) * 2018-06-07 2020-01-03 阿里巴巴集团控股有限公司 商品溯源链路信息处理方法、装置以及电子设备
CN111865771A (zh) * 2018-08-08 2020-10-30 创新先进技术有限公司 消息发送方法及装置和电子设备
CN110858134A (zh) * 2018-08-22 2020-03-03 阿里巴巴集团控股有限公司 数据、显示处理方法、装置、电子设备和存储介质
CN110858134B (zh) * 2018-08-22 2023-04-28 阿里巴巴集团控股有限公司 数据、显示处理方法、装置、电子设备和存储介质
CN111353839A (zh) * 2018-12-21 2020-06-30 阿里巴巴集团控股有限公司 商品信息处理方法、直播商品的方法、装置及电子设备
CN111353839B (zh) * 2018-12-21 2023-05-02 阿里巴巴集团控股有限公司 商品信息处理方法、直播商品的方法、装置及电子设备
CN111597863B (zh) * 2019-02-21 2023-11-28 顺丰科技有限公司 一种装卸率确定方法、系统、设备及存储介质
CN111597863A (zh) * 2019-02-21 2020-08-28 顺丰科技有限公司 一种装卸率确定方法、系统、设备及存储介质
US11012675B2 (en) 2019-04-16 2021-05-18 At&T Intellectual Property I, L.P. Automatic selection of viewpoint characteristics and trajectories in volumetric video presentations
US10970519B2 (en) 2019-04-16 2021-04-06 At&T Intellectual Property I, L.P. Validating objects in volumetric video presentations
US11074697B2 (en) 2019-04-16 2021-07-27 At&T Intellectual Property I, L.P. Selecting viewpoints for rendering in volumetric video presentations
US11153492B2 (en) 2019-04-16 2021-10-19 At&T Intellectual Property I, L.P. Selecting spectator viewpoints in volumetric video presentations of live events
US11470297B2 (en) 2019-04-16 2022-10-11 At&T Intellectual Property I, L.P. Automatic selection of viewpoint characteristics and trajectories in volumetric video presentations
US11663725B2 (en) 2019-04-16 2023-05-30 At&T Intellectual Property I, L.P. Selecting viewpoints for rendering in volumetric video presentations
US11670099B2 (en) 2019-04-16 2023-06-06 At&T Intellectual Property I, L.P. Validating objects in volumetric video presentations
US11956546B2 (en) 2019-04-16 2024-04-09 At&T Intellectual Property I, L.P. Selecting spectator viewpoints in volumetric video presentations of live events
CN111935488B (zh) * 2019-05-13 2022-10-28 阿里巴巴集团控股有限公司 数据处理方法、信息显示方法、装置、服务器及终端设备
CN111935488A (zh) * 2019-05-13 2020-11-13 阿里巴巴集团控股有限公司 数据处理方法、信息显示方法、装置、服务器及终端设备
CN110456901A (zh) * 2019-08-16 2019-11-15 上海电气集团股份有限公司 展会中对象展示的控制方法、系统、电子设备和存储介质
CN112132644A (zh) * 2020-08-21 2020-12-25 苏州合浩网络科技有限公司 一种vr商城的商品智能陈列方法及更新系统
CN112991553A (zh) * 2021-03-11 2021-06-18 深圳市慧鲤科技有限公司 信息展示方法及装置、电子设备和存储介质

Also Published As

Publication number Publication date
CN107633441A (zh) 2018-01-26

Similar Documents

Publication Publication Date Title
WO2018036456A1 (fr) Procédé et dispositif pour suivre et reconnaître une marchandise dans une image vidéo et afficher des informations de marchandise
US9930311B2 (en) System and method for annotating a video with advertising information
JP6952763B2 (ja) メディア表示と同期したコンテンツアイテムの提示
US11496814B2 (en) Method, system and computer program product for obtaining and displaying supplemental data about a displayed movie, show, event or video game
US11482192B2 (en) Automated object selection and placement for augmented reality
US20190095955A1 (en) Dynamic binding of live video content
KR101839927B1 (ko) 미디어 데이터에 있는 아이템을 인식하고 이와 관련된 정보를 전달하기 위한 시스템 및 방법
US11741681B2 (en) Interaction analysis systems and methods
CN110858134B (zh) 数据、显示处理方法、装置、电子设备和存储介质
EP3425483B1 (fr) Dispositif de reconnaissance d'objet intelligent
KR20160027486A (ko) 광고 제공 장치, 광고 표시 장치, 광고 제공 방법, 및 광고 표시 방법
CN107578306A (zh) 追踪识别视频图像中的商品并展示商品信息的方法和装置
CN112288877A (zh) 视频播放方法、装置、电子设备及存储介质
KR101573676B1 (ko) 메타데이터 기반의 객체기반 가상시점 방송 서비스 방법 및 이를 위한 기록매체
KR20140076674A (ko) 동영상에 객체 증강을 이용한 광고 시스템 및 방법

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17842879

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 03.06.2019)

122 Ep: pct application non-entry in european phase

Ref document number: 17842879

Country of ref document: EP

Kind code of ref document: A1