WO2020038167A1 - Procédé et appareil de reconnaissance d'image vidéo, terminal et support d'enregistrement - Google Patents

Procédé et appareil de reconnaissance d'image vidéo, terminal et support d'enregistrement Download PDF

Info

Publication number
WO2020038167A1
WO2020038167A1 PCT/CN2019/096578 CN2019096578W WO2020038167A1 WO 2020038167 A1 WO2020038167 A1 WO 2020038167A1 CN 2019096578 W CN2019096578 W CN 2019096578W WO 2020038167 A1 WO2020038167 A1 WO 2020038167A1
Authority
WO
WIPO (PCT)
Prior art keywords
image recognition
image
target
recognition result
identified
Prior art date
Application number
PCT/CN2019/096578
Other languages
English (en)
Chinese (zh)
Inventor
宋方
Original Assignee
Oppo广东移动通信有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Oppo广东移动通信有限公司 filed Critical Oppo广东移动通信有限公司
Publication of WO2020038167A1 publication Critical patent/WO2020038167A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content

Definitions

  • the embodiments of the present application relate to the technical field of terminals, and in particular, to a video image recognition method, device, terminal, and storage medium.
  • the terminal In related technology, if the user wants to know the characters or objects in the video, the terminal is usually triggered to take a screenshot of the current playback interface and save the screenshot, and then the terminal is triggered to exit the playback application and start running the search application. The user takes the above screenshot Upload to the search application and click the search control. At this time, the terminal obtains the relevant information about the person or item from the network and displays it to the user.
  • the embodiments of the present application provide a video image recognition method, device, terminal, and storage medium.
  • the technical solution is as follows:
  • an embodiment of the present application provides a video image recognition method, where the method includes:
  • the picture recognition control is displayed in the sidebar
  • the image recognition result is displayed.
  • an embodiment of the present application provides a video image recognition apparatus, where the apparatus includes:
  • a control display module is used to display the image recognition control in the sidebar when the video is playing.
  • An image acquisition module configured to, when receiving a first trigger signal corresponding to the image recognition function control, perform screenshot processing on a current playback screen to obtain a target image
  • An image recognition module configured to perform image recognition on the target image to obtain an image recognition result of the target image
  • a result display module configured to display the image recognition result.
  • an embodiment of the present application provides a terminal.
  • the terminal includes a processor and a memory.
  • the memory stores a computer program, and the computer program is loaded and executed by the processor to implement the method described in the foregoing aspect.
  • Video recognition method is another aspect.
  • an embodiment of the present application provides a computer-readable storage medium.
  • the computer-readable storage medium stores a computer program, and the computer program is loaded by a processor and executes the video image recognition according to the foregoing aspect. method.
  • FIG. 1 is a schematic structural diagram of a terminal provided by an exemplary embodiment of the present application.
  • FIG. 2 is a schematic structural diagram of a terminal according to another exemplary embodiment of the present application.
  • 3A to 3F are schematic diagrams of appearances of terminals with different touch display screens provided by exemplary embodiments of the present application.
  • FIG. 4 is a flowchart of a video image recognition method according to an embodiment of the present application.
  • FIG. 5 is a schematic diagram of an interface for displaying a video image recognition control provided by an embodiment of the present application
  • FIG. 6 is a schematic diagram of an interface for determining a target to-be-identified object according to an embodiment of the present application
  • FIG. 7 is a schematic diagram of an interface for determining a target to-be-recognized object according to another embodiment of the present application.
  • FIG. 8 is a schematic diagram of an interface for determining a target to-be-recognized object according to another embodiment of the present application.
  • FIG. 9 is a schematic interface diagram of a video image recognition method provided by an embodiment of the present application.
  • FIG. 10 is a schematic interface diagram of a video image recognition method provided by an embodiment of the present application.
  • FIG. 11 is a schematic diagram of an interface where video image recognition fails according to an embodiment of the present application.
  • FIG. 12 is a schematic diagram of an interface where video image recognition fails according to an embodiment of the present application.
  • FIG. 13 is a block diagram of a video image recognition apparatus according to an embodiment of the present application.
  • FIG. 1 and FIG. 2 are structural block diagrams of a terminal 100 according to an exemplary embodiment of the present application.
  • the terminal 100 may be a mobile phone, a tablet computer, a notebook computer, an e-book, or the like.
  • the terminal 100 in the present application may include one or more of the following components: a processor 110, a memory 120, and a touch display screen 130.
  • the processor 110 may include one or more processing cores.
  • the processor 110 uses various interfaces and lines to connect various parts in the entire terminal 100, and executes the terminal by running or executing instructions, programs, code sets or instruction sets stored in the memory 120, and calling data stored in the memory 120. 100 various functions and processing data.
  • the processor 110 may use at least one of Digital Signal Processing (DSP), Field-Programmable Gate Array (FPGA), and Programmable Logic Array (PLA).
  • DSP Digital Signal Processing
  • FPGA Field-Programmable Gate Array
  • PDA Programmable Logic Array
  • the processor 110 may integrate one or a combination of a central processing unit (Central Processing Unit, CPU), an image processor (Graphics Processing Unit, GPU), and a modem.
  • CPU Central Processing Unit
  • GPU Graphics Processing Unit
  • modem modem
  • the CPU mainly handles the operating system, user interface, and application programs; the GPU is responsible for rendering and rendering of the content required to be touched by the display screen 130; and the modem is used for processing wireless communication. It can be understood that the modem may not be integrated into the processor 110, and may be implemented by a single chip.
  • the memory 120 may include Random Access Memory (RAM), and may also include Read-Only Memory.
  • the memory 120 includes a non-transitory computer-readable storage medium.
  • the memory 120 may be used to store instructions, programs, codes, code sets, or instruction sets.
  • the memory 120 may include a storage program area and a storage data area, where the storage program area may store instructions for implementing an operating system, instructions for at least one function (such as a touch function, a sound playback function, an image playback function, etc.), Instructions for implementing the following method embodiments, etc .; the storage data area may store data (such as audio data, phone book) and the like created according to the use of the terminal 100.
  • the memory 120 stores a Linux kernel layer 220, a system runtime layer 240, an application framework layer 260, and an application layer 280.
  • the Linux kernel layer 220 provides low-level drivers for various hardware of the terminal 100, such as a display driver, an audio driver, a camera driver, a Bluetooth driver, a Wi-Fi driver, and power management.
  • the system runtime layer 240 provides main feature support for the Android system through some C / C ++ libraries. For example, the SQLite library provides database support, the OpenGL / ES library provides 3D graphics support, and the Webkit library provides browser kernel support.
  • An Android runtime library 242 (Android Runtime) is also provided in the system runtime layer 240, which mainly provides some core libraries, which can allow developers to write Android applications using the Java language.
  • the application framework layer 260 provides various APIs that may be used when building applications. Developers can also use these APIs to build their own applications, such as activity management, window management, view management, notification management, content providers, Package management, call management, resource management, positioning management.
  • At least one application program is running in the application layer 280, and these application programs may be contact programs, SMS programs, clock programs, camera applications, etc. that are native to the operating system; or applications developed by third-party developers, such as instant Communication programs, photo beautification programs, etc.
  • the IOS system includes: a core operating system layer 320 (Core OS layer), a core service layer 340 (Core services layer), and a media layer 360 (Media layer), 380 (Cocoa Touch Layer).
  • the core operating system layer 320 includes an operating system kernel, drivers, and a low-level program framework. These low-level program frameworks provide functions closer to the hardware for use by the program framework located in the core service layer 340.
  • the core service layer 340 provides system services and / or program frameworks required by applications, such as a Foundation framework, an account framework, an advertising framework, a data storage framework, a network connection framework, a geographic location framework, a motion framework, and so on.
  • the media layer 360 provides audio-visual-related interfaces for applications, such as interfaces related to graphics and images, interfaces related to audio technology, interfaces related to video technology, and wireless playback (AirPlay) interfaces for audio and video transmission technologies.
  • the touchable layer 380 provides various commonly-used interface-related frameworks for application development. The touchable layer 380 is responsible for user touch interaction operations on the terminal 100. For example, a local notification service, a remote push service, an advertising framework, a game tool framework, a message user interface interface (UI) framework, a user interface UIKit framework, a map framework, and so on.
  • UI message user interface interface
  • frameworks related to most applications include, but are not limited to, a basic framework in a core service layer 340 and a UIKit framework in a touchable layer 380.
  • the basic framework provides many basic object classes and data types, and provides the most basic system services for all applications, regardless of the UI.
  • the classes provided by the UIKit framework are basic UI class libraries for creating touch-based user interfaces.
  • IOS applications can provide UIs based on the UIKit framework, so it provides the application's infrastructure for building user interfaces and drawing. , Handle events with user interaction, respond to gestures, and more.
  • the touch display screen 130 is used for receiving a user's touch operation on or near any suitable object using a finger, a touch pen, or the like, and displaying a user interface of each application program.
  • the touch display screen 130 is generally disposed on a front panel of the terminal 130.
  • the touch display screen 130 may be designed as a full screen, a curved screen or a special-shaped screen.
  • the touch display screen 130 can also be designed as a combination of a full screen and a curved screen, and a combination of a special-shaped screen and a curved screen, which is not limited in this embodiment. among them:
  • the full screen may refer to a screen design in which the touch screen display 130 occupies the front panel of the terminal 100 with a screen ratio exceeding a threshold (such as 80% or 90% or 95%).
  • One calculation method of the screen ratio is: (the area of the touch display screen 130 / the area of the front panel of the terminal 100) * 100%; another calculation method of the screen ratio is: (the actual display area in the touch display screen 130) Area / the area of the front panel of the terminal 100) * 100%; another way to calculate the screen ratio is: (the diagonal of the touch screen 130 / the diagonal of the front panel of the terminal 100) * 100% .
  • a threshold such as 80% or 90% or 95%).
  • One calculation method of the screen ratio is: (the area of the touch display screen 130 / the area of the front panel of the terminal 100) * 100%; another calculation method of the screen ratio is: (the actual display area in the touch display screen 130) Area / the area of the front panel of the terminal 100) * 100%; another way to calculate the screen ratio is: (the
  • the full screen may also be a screen design in which at least one front panel component is integrated inside or below the touch display screen 130.
  • the at least one front panel component includes a camera, a fingerprint sensor, a proximity light sensor, a distance sensor, and the like.
  • other components on the front panel of the conventional terminal are integrated in all or part of the touch display screen 130. For example, after the photosensitive element in the camera is divided into multiple photosensitive pixels, each photosensitive The pixels are integrated in a black area in each display pixel in the touch display screen 130. Since at least one front panel component is integrated inside the touch display screen 130, the full screen has a higher screen ratio.
  • the front panel components on the front panel of the traditional terminal can also be set on the side or back of the terminal 100.
  • an ultrasonic fingerprint sensor is set under the touch display screen 130, and a bone conduction type
  • the handset is disposed inside the terminal 130, and the camera is disposed on the side of the terminal and is pluggable.
  • a single side of the middle frame of the terminal 100 or two sides (such as left and right sides), or four sides (such as (Upper, lower, left, and right sides) are provided with edge touch sensors 120, which are used to detect the user's touch operations, click operations, press operations, and slide operations on the middle frame. At least one operation.
  • the edge touch sensor 120 may be any one of a touch sensor, a thermal sensor, and a pressure sensor. The user can apply an operation on the edge touch sensor 120 to control an application program in the terminal 100.
  • the curved screen refers to a screen design in which the cross-section of the touch display screen 130 has a curved shape and the projection is a plane in a direction parallel to the cross-section.
  • the curved shape may be U-shaped.
  • the curved screen refers to a screen design manner in which at least one side is a curved shape.
  • the curved screen refers to that at least one side of the touch display screen 130 extends to cover the middle frame of the terminal 100. Since the side of the touch display screen 130 extends to the middle frame of the terminal 100, the middle frame without the display function and the touch function is also covered as a displayable area and / or an operable area, so that the curved screen has a more High screen ratio.
  • the curved screen refers to a screen design in which the left and right sides 42 are curved; or the curved screen refers to a screen design in which the upper and lower sides are curved; or, Curved screen refers to a screen design with four curved sides on the top, bottom, left, and right.
  • the curved screen is made of a touch screen material with a certain flexibility.
  • the special-shaped screen is a touch display screen with an irregular appearance.
  • the irregular shape is not a rectangle or a rounded rectangle.
  • the special-shaped screen refers to a screen design provided with protrusions, notches, and / or holes on the rectangular or rounded rectangular touch display screen 130.
  • the protrusion, the notch and / or the hole can be located at the edge of the touch display screen 130, the center of the screen, or both.
  • the protrusion, notch and / or hole When the protrusion, notch and / or hole is set on one edge, it can be set at the middle position or both ends of the edge; when the protrusion, notch and / or hole is set on the center of the screen, it can be set above the screen One or more of the region, the upper left region, the left region, the lower left region, the lower region, the lower right region, the right region, and the upper right region.
  • the protrusions, notches, and digging holes can be distributed in a centralized or distributed manner; they can be distributed symmetrically or asymmetrically.
  • the number of the protrusions, notches and / or holes is not limited.
  • the special-shaped screen covers the upper and / or lower forehead area of the touch display screen as a displayable area and / or an operable area, so that the touch-screen display takes up more space on the front panel of the terminal. Larger screen ratio.
  • the notches and / or holes are used to receive at least one front panel component, which includes a camera, a fingerprint sensor, a proximity light sensor, a distance sensor, a handset, an ambient light sensor, and a physical button. At least one.
  • the notch may be provided on one or more edges, and the notch may be a semicircular notch, a right-angled rectangular notch, a rounded rectangular notch, or an irregularly shaped notch.
  • the special-shaped screen may be a screen design provided with a semi-circular notch 43 at the center of the upper edge of the touch display screen 130, and the space vacated by the semi-circular notch 43 is used.
  • the special-shaped screen may be on the lower edge of the touch display screen 130
  • the screen design of the semi-circular notch 44 is set at the center position, and the space vacated by the semi-circular notch 44 is used to accommodate at least one component of a physical button, a fingerprint sensor, and a microphone; as shown schematically in FIG. 3E
  • the special-shaped screen may be a screen design provided with a semi-elliptical notch 45 at the center of the lower edge of the touch display screen 130, and a semi-elliptical notch is formed on the front panel of the terminal 100.
  • the notch encloses an elliptical area, which is used to accommodate physical keys or fingerprint recognition modules.
  • the special-shaped screen can be touched.
  • the upper half of the display screen 130 is provided with a screen design of at least one small hole 46.
  • the space vacated by the small hole 46 is used to accommodate at least one of the front panel of the camera, distance sensor, handset, and ambient light sensor. component.
  • the structure of the terminal 100 shown in the above drawings does not constitute a limitation on the terminal 100.
  • the terminal may include more or fewer components than shown, or a combination of some Components, or different component arrangements.
  • the terminal 100 further includes components such as a radio frequency circuit, an input unit, a sensor, an audio circuit, a wireless fidelity (WiFi) module, a power source, and a Bluetooth module, and details are not described herein again.
  • WiFi wireless fidelity
  • embodiments of the present application provide a method, a device, a terminal, and a storage medium for video image recognition.
  • the terminal displays a video recognition control in a video playback scene. If the user desires to know a certain character or item in the current playback screen, directly click the video recognition control, and then the terminal displays the current playback The screen performs image recognition and displays the image recognition results to the user. This process can prevent the user from switching back and forth between the two applications, saving the user the operation required to understand a character or item in the currently playing screen, and improving work efficiency. .
  • the execution subject of each step may be a terminal described in the foregoing embodiment.
  • the terminal has a video playing function.
  • the terminal also has an image recognition function.
  • an application for implementing a video playback function is installed and run in the terminal, and the execution subject of each step may be the application, and the application may be a system application or a third-party application.
  • the execution subject of each step is a terminal is used as an example for description, but this is not a limitation.
  • FIG. 4 shows a flowchart of a video image recognition method provided by an embodiment of the present application.
  • the method may include the following steps:
  • Step 401 When in a video playback scene, display a picture recognition function control in a sidebar.
  • a video playing scene refers to a scene where a terminal is playing a video.
  • the terminal plays a video by playing an application program; in another possible implementation manner, the terminal plays a video in a webpage through a browser.
  • the sidebar is used to display application icons and / or function controls in the terminal, so that the terminal can conveniently open other applications or execute functions corresponding to the function controls while the terminal is running the application in the foreground.
  • the application icons and / or function controls displayed in the sidebar can be set by the terminal by default or can be customized by the user.
  • the image recognition function control is included in the sidebar.
  • the image recognition function control is used to trigger image recognition of the picture in the currently playing video.
  • the identification function control may be displayed when the video starts to play, or may be displayed according to an operation signal triggered by a user.
  • the embodiment of the present application does not limit the display timing of the identification function control.
  • step 401 may include the following two sub-steps:
  • Step 401a when in a video playing scene, receive an outgoing call instruction corresponding to a sidebar;
  • step 401b a sidebar is displayed according to the outgoing call instruction.
  • the call out command is used to call out the sidebar.
  • a buoy is displayed on the display interface of the terminal, and if a trigger signal acting on the buoy is received, the terminal receives an outgoing call instruction.
  • the buoy can always be displayed on the upper layer of the display interface, or can be displayed on the upper layer of the display interface when the application is started and run, and can also be displayed on the upper layer of the display interface according to the operation signal triggered by the user.
  • the embodiment of the present application does not limit the display timing of the buoy.
  • the shape of the buoy may be a circle, an oval, a rectangle, or the like, and the shape of the buoy is not limited in the embodiment of the present application.
  • the area of the buoy can be set by the terminal by default, or can be set by the user, which is not limited in the embodiment of the present application.
  • the buoy in order to reduce the occlusion of the display interface as much as possible, the buoy can be set to a transparency greater than 0.
  • the trigger signal acting on the buoy may be any one of a click signal, a double-click signal, a long-press signal, a slide signal, and a drag signal, which is not limited in the embodiment of the present application.
  • the trigger signal acting on the buoy is a slide signal as an example for description.
  • the terminal when the terminal is in the horizontal screen display state, the buoy will block the display interface, and the user's immersion is reduced at this time.
  • the terminal when the terminal receives a trigger signal on the display side, it receives an outgoing call instruction.
  • the trigger signal on the display side is a sliding signal from the outside to the inside of the display side.
  • FIG. 5 a schematic diagram of an interface for displaying a video image recognition control according to an embodiment of the present application is shown.
  • the user When in the video playback scene, the user performs a sliding operation from the outside to the inside of the left side of the terminal. After receiving the sliding operation signal, the terminal displays a sidebar 51, and the sidebar 51 includes a picture recognition control 52.
  • step 402 when a first trigger signal corresponding to the image recognition function control is received, screenshot processing is performed on the current playback screen to obtain a target image.
  • the first trigger signal is triggered by the user, and may be any one of a click signal, a double-click signal, a long-press signal, a slide signal, and a drag signal.
  • the first trigger signal is a click signal as an example for description.
  • the target image is an image that needs to be identified.
  • the terminal determines the playback picture displayed when the first trigger signal is received as the target image.
  • the target image needs to be displayed to the user, so that the user determines whether it is an image that needs to be identified.
  • the target image is acquired by a screenshot processing method. Screenshot processing refers to capturing the current playback frame and determining the captured playback frame as the target image.
  • the terminal performs screenshot processing on a complete current playback picture to obtain a target image.
  • the terminal performs a screenshot process on a part of the pictures currently being played to obtain a target image.
  • Some of the above screens can be selected by the user.
  • the video playback is paused and the user is prompted to capture the target image, the user performs a drag operation on the current playback screen, and then the terminal intercepts the diagonal line as the drag A rectangular area of a straight line from the start point to the end point of the operation signal is used as the target image.
  • the terminal After the terminal acquires the target image, it can also display the target image. Optionally, the terminal displays the target image on a floating window. Because the size of the floating window is small, when the target image is displayed on the floating window, the target image needs to be reduced in size.
  • Step 403 Obtain an image recognition result of the target image.
  • the image recognition result is obtained by performing image recognition on the target image.
  • the image recognition result may include at least one record, and each record is used to represent a recognition result of an element in the target image, which may be a person identification or an item identification.
  • the character identification is used to uniquely identify the person, and it may be the name of the person.
  • the terminal recognizes the person in the current playback screen and obtains the character identification.
  • the item identifier is used to uniquely identify the item, which may be the name of the item.
  • the terminal recognizes the item in the currently playing screen and obtains the item identification.
  • the image recognition result also includes the similarity corresponding to each record.
  • the similarity refers to the similarity between the record and the corresponding element in the target image, and is used to measure the accuracy of the image recognition result. The higher the similarity, the more accurate the image recognition result; the lower the similarity, the less accurate the image recognition result is.
  • the terminal recognizes a target image to obtain an image recognition result.
  • the server performs image recognition on the target image to obtain an image recognition result, and then the terminal obtains the image recognition result from the server.
  • the terminal sends an identification request to the server, and the identification request carries the identification of the terminal and the target image, and the server recognizes the target image according to the identification request, obtains an image recognition result, and returns the image recognition result to the terminal.
  • the terminal sends an identification request to the server, and the identification request carries the identification of the terminal and the target image, and the server recognizes the target image according to the identification request, obtains an image recognition result, and returns the image recognition result to the terminal.
  • the embodiment of the present application does not limit the algorithm used for image recognition. It may be an image recognition algorithm based on model matching, an image recognition algorithm based on neural networks, an image recognition algorithm based on wavelet moments, an image recognition algorithm based on fractal features, and so on. This is not limited in the embodiments of the present application.
  • the terminal may further display query information, where the query information is used to query whether it is necessary to obtain an image recognition result of the target image.
  • the terminal may further display query information, where the query information is used to query whether it is necessary to obtain an image recognition result of the target image.
  • Step 404 Display the image recognition result.
  • the terminal After the terminal obtains the image recognition result, the terminal displays the image recognition result for the user to view. Optionally, the image recognition result is also displayed in the floating window mentioned in step 402.
  • the technical solution provided in the embodiments of the present application by displaying a video recognition control in a video playback scene, if the user desires to know a certain character or item in the current playback screen, directly click the video recognition control, and then the terminal Perform image recognition on the current playback screen and display the image recognition results to the user.
  • This process can prevent users from switching back and forth between the two applications, saving the user the operation required to understand a character or item in the current playback screen. The operation is more convenient and the recognition efficiency is higher.
  • an image may include multiple objects, such as people, objects, animals, flowers, trees, etc.
  • the terminal still performs image recognition on the entire image, it may result in low recognition efficiency.
  • the user selects an object to be identified among the multiple objects, and then the terminal obtains only the image recognition result of the object to be recognized, without acquiring the image recognition result of the entire image, which can improve the recognition efficiency.
  • the target image includes multiple objects to be identified, and step 403 includes the following two sub-steps:
  • Step 501 Determine a target to-be-recognized object included in the target image.
  • the target to-be-recognized object refers to an object that the user desires to recognize, which can be selected by the user.
  • the number of target to-be-recognized objects may be one or multiple.
  • the number of target to-be-recognized objects may be less than the number of objects contained in the target image, or may be equal to the number of objects contained in the target image.
  • step 501 includes the following sub-steps:
  • Step 501a displaying a person identification control and / or an item identification control
  • the person recognition control is used to trigger the recognition of the area containing the person image in the target image
  • the item recognition control is used to trigger the recognition of the area containing the object image in the target image.
  • the terminal displays the above-mentioned person recognition control and / or item recognition control while displaying the target image.
  • the above-mentioned person recognition control and / or item recognition control are also displayed in the floating window.
  • step 501b when a second trigger signal corresponding to the person recognition control is received, it is determined that the target object to be identified is an area including a person image in the target image;
  • the area containing the person image in the target image may be a rectangular area containing a face image. Further, the area containing the person image in the target image is the smallest rectangular area containing the face image.
  • step 501c when a third trigger signal corresponding to the item identification control is received, it is determined that the target object to be identified is an area in the target image that includes the item image.
  • the area containing the image of the article in the target image may be the area containing the entire article or a rectangular area containing the key features of the article.
  • the key characteristics of the item can be determined based on the actual item. For example, when the item is a flower, its key feature is a petal. Further, the area containing the image of the article in the target image may be the smallest rectangular area containing the entire article, or the smallest rectangular area containing key features of the article.
  • FIG. 6 a schematic diagram of an interface for determining an object to be identified is provided according to an embodiment of the present application.
  • the terminal displays a target image 62, a person recognition control 63, and an article recognition control 64 on the floating window 61.
  • the terminal determines that the target to-be-recognized object is a region containing the person image in the target image;
  • the terminal determines that the target to-be-recognized object is an area that includes the object image in the target image.
  • step 501 includes the following sub-steps:
  • Step 501d displaying a target image
  • Each object to be identified in the target image is labeled with a different serial number.
  • the terminal also displays the above-mentioned different serial numbers below the target image.
  • Step 501e Receive a selection signal corresponding to the target sequence number
  • the selection signal corresponding to the target number may be any one of a click signal, a double-click signal, a long-press signal, a slide signal, and a drag signal, which is not limited in this embodiment of the present application.
  • the selection signal corresponding to the target sequence number is a click signal as an example for description.
  • the target sequence number is the selected sequence number. If the user wants to know about an object, he can select the serial number corresponding to the object. If the terminal also displays the different serial numbers above the target image, the terminal may select the target serial number in the target image, or select the target serial number among the serial numbers displayed below the target image.
  • step 501f the object to be identified corresponding to the target number is determined as the target object to be identified.
  • the terminal determines the object corresponding to the selected serial number as the target object to be identified.
  • the floating window further includes a completion control, and when the terminal receives a confirmation instruction corresponding to the completion control, the object corresponding to the selected serial number is determined as the object to be identified.
  • FIG. 7 a schematic diagram of an interface for determining an object to be identified is provided according to an embodiment of the present application.
  • the terminal displays a target image 62 and a completion control 71 on the floating window 61.
  • Each to-be-recognized object in the target image 62 is identified with a different serial number.
  • the terminal When the user clicks a certain serial number and the completion control 71, the terminal corresponding to the serial number
  • the recognition object is determined as the target to-be-recognized object.
  • step 501 includes the following sub-steps:
  • Step 501g displaying a target image
  • Step 501h receiving a third trigger signal acting on the target image
  • the third trigger signal may be any one of a click signal, a double-click signal, a long-press signal, a slide signal, and a drag signal, which is not limited in the embodiment of the present application.
  • the object to be identified in the target area corresponding to the third trigger signal is determined as the target object to be identified.
  • the target area corresponding to the third trigger signal refers to the trigger position of the third trigger signal as the center and the area is a preset area Area.
  • the trigger position of the third trigger signal refers to a contact position between the user's finger and the display screen.
  • the preset area can be set according to actual experience, which is not limited in the embodiment of the present application.
  • the target area corresponding to the third trigger signal is a rectangular area with the motion track of the third trigger signal as a diagonal.
  • the floating window further includes a completion control, and when the terminal receives a confirmation instruction corresponding to the completion control, the object in the target area corresponding to the third trigger signal is determined as the target object to be identified.
  • FIG. 8 a schematic diagram of an interface for determining an object to be identified is provided according to another embodiment of the present application.
  • the terminal displays a target image 62 on the floating window 61.
  • the terminal determines the target object to be identified in the area 81 centered on the position and having an area of a preset area.
  • Step 302 Perform image recognition on the target to-be-recognized object to obtain an image recognition result.
  • step 302 may be implemented as: performing image recognition on the target object to be recognized through a machine learning model to obtain an image recognition result.
  • a machine learning model is obtained by training a neural network using multiple sets of training sample data.
  • Each set of training sample data in the plurality of sets of training sample data includes a sample image and a recognition result corresponding to the sample image.
  • the recognition result corresponding to the sample image can be obtained manually, that is, the relevant technician determines the recognition result corresponding to the sample image and records it.
  • the neural network may be a Convolutional Neural Network (CNN), an Artificial Neural Network (ANN), a Deep Neural Networks (DNN), and the like, which are not limited in the embodiments of the present application.
  • CNN Convolutional Neural Network
  • ANN Artificial Neural Network
  • DNN Deep Neural Networks
  • the machine learning algorithm used in training the machine learning model can be a back-propagation (BP) algorithm, a faster regional convolutional neural network faster RCNN (faster Regions with Convolutional Neural Network, faster RCNN) algorithm, etc., this application
  • BP back-propagation
  • RCNN faster Regions with Convolutional Neural Network
  • the machine learning model includes: an input layer, at least one hidden layer, and an output layer.
  • the input data of the input layer is the target image or the target object to be identified in the target image
  • the output result of the output layer is the image recognition result of the target image.
  • the determination process is as follows: the target image or the object to be identified in the target image is input to the input layer of the machine learning model, and the hidden layer of the machine learning model performs feature extraction on the above feature data, and combines and abstracts the extracted features , And finally the image recognition result of the target image is output by the output layer.
  • the specific structure of the hidden layer is not limited. Generally speaking, the more layers of a neural network, the better the effect but the longer the calculation time. In practical applications, a neural network with an appropriate number of layers can be designed in accordance with the accuracy requirements.
  • the training process of the machine learning model is as follows: obtaining the initial machine learning model, inputting the sample images in the sample training data to the initial machine learning model, and outputting the actual recognition results corresponding to the sample images from the initial machine learning model, The recognition result is compared with the image recognition result corresponding to the sample image to obtain the calculated loss, and then the calculated loss is compared with a preset threshold. If the calculated loss is greater than the preset threshold, the parameters of the initial machine learning model are updated. Then, the steps of inputting the sample images in the sample training data to the initial machine learning model are restarted. If the calculation loss is not greater than a preset threshold, a machine learning model is generated.
  • the preset threshold may be determined actually according to the recognition accuracy, which is not limited in the embodiment of the present application.
  • the technical solution provided in the embodiments of the present application allows the user to first identify the person or article to be identified in the image to be identified, and does not need to perform image recognition on the entire image during subsequent image recognition.
  • the selected people or objects for identification can improve the efficiency of image recognition.
  • the terminal may also obtain and display related information corresponding to the image recognition result, so that the user can know more abundant and comprehensive information about the person or article in the playback screen.
  • the video image recognition method may further include the following steps:
  • Step 601 Obtain related information corresponding to the image recognition result.
  • the related information corresponding to the image recognition result includes one or more of the following combinations: encyclopedia information, social account information, news information information, and work information of the person corresponding to the person identification.
  • Encyclopedia information refers to the detailed information of the person, which usually includes name, age, occupation, birthday, and so on.
  • the social account information includes a web page link of the social account used by the character. When the web page link is clicked, the terminal displays the main page of the social account so that the user can establish a social relationship with the social account by using the social account. It can be concern relationship, listening relationship, friend relationship, etc.
  • News information refers to news information related to the person.
  • the work information includes a detailed introduction to the work in which the character has appeared, and a link to visit.
  • the related information corresponding to the image recognition result includes one or more of the following combinations: encyclopedia information and purchase information of the item corresponding to the item identification.
  • Encyclopedia information refers to the detailed information of the item, which can include the name, material, weight, etc. of the item.
  • the purchase information includes a purchase link for the item. When the purchase link is clicked, the terminal displays a purchase page for the item so that the user can purchase the item.
  • the terminal acquires the related information of the image recognition result locally.
  • the terminal obtains related information of the image recognition result from the server. Specifically, the terminal sends an acquisition request to the server, and the identification request carries the identification of the terminal and the image recognition result. The server obtains related information corresponding to the image recognition result according to the acquisition request, and returns the related information to the terminal.
  • the second possible implementation manner is used as an example for explanation.
  • Step 602 Display related information corresponding to the image recognition result.
  • the terminal directly jumps to display related information corresponding to the image recognition result.
  • the terminal displays a jump control while displaying the image recognition result, and when the terminal receives a trigger signal corresponding to the jump control, displays related information corresponding to the image recognition result.
  • the terminal displays the jump control corresponding to each record.
  • the terminal displays related information of the record corresponding to the target jump control.
  • the terminal when the terminal displays related information corresponding to the image recognition result, it also displays a favorite control.
  • the terminal receives the trigger signal corresponding to the favorite control, the terminal saves relevant information corresponding to the image recognition result.
  • the favorite control will change to the favorited state.
  • the terminal directly stores the foregoing related information in the first storage path, and subsequent users can directly view the related information when there is no network connection, thereby reducing traffic consumption.
  • the terminal stores the access address corresponding to the related information in the second storage path, and subsequent users can obtain and view related information again through the access address, thereby reducing the storage space occupation of the terminal.
  • the first storage path and the second storage path may be set by a user, or may be set by a terminal by default, which is not limited in the embodiment of the present application.
  • each item corresponds to a favorite control, so that the user can selectively save the related related information that he needs.
  • FIG. 9 illustrates a schematic diagram of an interface for displaying related information provided by an embodiment of the present application.
  • the terminal displays the target image 62, the person recognition control 63, and the object recognition control 64 in the floating window 61.
  • the terminal displays the first record 91 "person in the image recognition result" in the floating window 61 A ", jump control 92 corresponding to the first record 91" Person A “, jump control 94 corresponding to the second record 93" Person B “and second record 93" Person B “; when the user clicks on the first
  • the jump control 92 corresponding to the record 91 "Person A the floating window 61 displays the related information 95 corresponding to the first record 91 "Person A” and the favorite control 96.
  • FIG. 10 a schematic diagram of an interface for displaying related information provided by an embodiment of the present application is shown.
  • the terminal displays a target image 62 in the floating window 61.
  • the terminal determines an object in the area 1001 centered on the position and having an area of a preset area as the object to be identified, and then the terminal obtains the object's Image recognition results.
  • the image recognition results include a record "authentic baseball caps, tide brand hats, sun hats, men and women.”
  • the terminal directly displays a plurality of related information 1002 of the image recognition results in the floating window 61, and the corresponding collection of each related information Control 1003.
  • the terminal When the terminal obtains the image recognition result, there may be a case where the image recognition result is not obtained.
  • the terminal displays first prompt information, and the first prompt information is used to prompt that relevant information cannot be obtained.
  • FIG. 11 a schematic diagram of an interface of the first prompt information provided by an embodiment of the present application is shown.
  • the terminal fails to obtain the image recognition result, the terminal displays the target image 62 and the first prompt information 1101 “No relevant information found” in the floating window 61.
  • the terminal displays a second prompt message at this time, and the second prompt information is used to prompt the user to establish a network connection, so that the terminal can retrieve relevant information.
  • the terminal also displays a network setting control. When the terminal receives a trigger signal corresponding to the network setting control, it jumps to the network setting interface so that the user can complete the network setting.
  • FIG. 12 a schematic diagram of an interface of the second prompt information provided by one embodiment of the present application is shown.
  • the target image 62 and the first prompt message 1201 “Please try again after connecting to the network” and the network setting control 1202 are displayed in the floating window 61.
  • FIG. 13 is a block diagram of a video image recognition apparatus provided by an embodiment of the present application.
  • the device has a function for implementing the above method example, and the function may be implemented by hardware, or may be implemented by hardware executing corresponding software.
  • the device may include:
  • the control display module 1301 is configured to display a picture recognition function control in a sidebar when the video is playing.
  • the image acquisition module 1302 is configured to perform a screenshot process on a current playback screen when a first trigger signal corresponding to the image recognition function control is received, to obtain a target image.
  • An image recognition module 1303 is configured to obtain an image recognition result of the target image, where the image recognition result is obtained by performing image recognition on the target image.
  • a result display module 1304 is configured to display the image recognition result.
  • the technical solution provided in the embodiments of the present application by displaying a video recognition control in a video playback scene, if the user desires to know a certain character or item in the current playback screen, directly click the video recognition control, and then the terminal Perform image recognition on the current playback screen and display the image recognition results to the user.
  • This process can prevent users from switching back and forth between the two applications, saving the user the operation required to understand a character or item in the current playback screen. The operation is more convenient and the recognition efficiency is higher.
  • the target image includes multiple objects to be identified, and the image recognition module 1303 is configured to:
  • the image recognition module 1303 is configured to:
  • determining that the target object to be identified is a region in the target image that includes a person image
  • the target object to be identified is an area in the target image that includes an item image.
  • the image recognition module 1303 is configured to:
  • the object to be identified corresponding to the target sequence number is determined as the object to be identified.
  • the image recognition module 1303 is configured to:
  • An object to be identified in a target area corresponding to the third trigger signal is determined as the target object to be identified.
  • the image recognition module 1303 is configured to perform image recognition on the target to-be-recognized object through a machine learning model to obtain the image recognition result.
  • the machine learning model is to use multiple sets of training sample data to the neural network. According to training, each set of training sample data in the plurality of sets of training sample data includes a sample image and a recognition result corresponding to the sample image.
  • the device further includes: an information acquisition module and an information display module (not shown in the figure)
  • An information acquisition module is configured to acquire related information corresponding to the image recognition result.
  • An information display module is configured to display related information corresponding to the image recognition result.
  • the related information corresponding to the image recognition result includes one or more of the following combinations: encyclopedia information, social account information, news information information, works of the person corresponding to the person identification information;
  • the related information corresponding to the image recognition result includes one or more of the following combinations: encyclopedia information and purchase information of the item corresponding to the item identification.
  • the information display module is configured to:
  • control display module 1301 is configured to:
  • a computer-readable storage medium stores a computer program, and the computer program is loaded and executed by a processor of a terminal to implement the foregoing method embodiments. Steps.
  • a computer program product is also provided, and when the computer program product is executed, it is used to implement the functions of each step in the foregoing method embodiments.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

Les modes de réalisation de la présente invention concernent un procédé et un appareil de reconnaissance d'image vidéo, un terminal et un support d'enregistrement. Le procédé comprend les étapes consistant à : afficher, dans une scène de lecture vidéo, une commande de fonction de reconnaissance d'image dans une barre latérale ; effectuer, lorsqu'un premier signal de déclenchement correspondant à la commande de fonction de reconnaissance d'image est reçu, un traitement de capture d'écran sur l'image actuellement lue pour obtenir une image cible ; obtenir un résultat de reconnaissance d'image de l'image cible ; et afficher le résultat de reconnaissance d'image. Dans les modes de réalisation de la présente invention, une commande de reconnaissance d'image vidéo est affichée dans la scène de lecture vidéo, si un utilisateur souhaite connaître une personne ou un élément dans l'image actuellement lue, l'utilisateur clique directement sur la commande de reconnaissance vidéo, et ensuite, le terminal effectue une reconnaissance d'image sur l'image actuellement lue et affiche le résultat de reconnaissance d'image à l'utilisateur. Grâce à ce procédé, l'utilisateur n'a pas besoin de basculer entre deux programmes d'application, l'opération requise par l'utilisateur pour connaître une personne ou un élément dans l'image actuellement lue est simplifiée, l'opération est plus pratique, et l'efficacité de reconnaissance d'image est plus élevée.
PCT/CN2019/096578 2018-08-22 2019-07-18 Procédé et appareil de reconnaissance d'image vidéo, terminal et support d'enregistrement WO2020038167A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201810963246.7A CN109034115B (zh) 2018-08-22 2018-08-22 视频识图方法、装置、终端及存储介质
CN201810963246.7 2018-08-22

Publications (1)

Publication Number Publication Date
WO2020038167A1 true WO2020038167A1 (fr) 2020-02-27

Family

ID=64628027

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/096578 WO2020038167A1 (fr) 2018-08-22 2019-07-18 Procédé et appareil de reconnaissance d'image vidéo, terminal et support d'enregistrement

Country Status (2)

Country Link
CN (1) CN109034115B (fr)
WO (1) WO2020038167A1 (fr)

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111444822A (zh) * 2020-03-24 2020-07-24 北京奇艺世纪科技有限公司 对象识别方法和装置、存储介质和电子装置
CN111541907A (zh) * 2020-04-23 2020-08-14 腾讯科技(深圳)有限公司 物品显示方法、装置、设备及存储介质
CN112565863A (zh) * 2020-11-26 2021-03-26 深圳Tcl新技术有限公司 视频播放方法、装置、终端设备及计算机可读存储介质
CN112584213A (zh) * 2020-12-11 2021-03-30 海信视像科技股份有限公司 一种显示设备和图像识别结果的展示方法
CN112801004A (zh) * 2021-02-05 2021-05-14 网易(杭州)网络有限公司 视频片段的筛选方法、装置、设备及存储介质
CN113747182A (zh) * 2021-01-18 2021-12-03 北京京东拓先科技有限公司 物品展示方法、客户端、直播服务器和计算机存储介质
CN113766297A (zh) * 2021-05-27 2021-12-07 腾讯科技(深圳)有限公司 视频处理方法、播放终端及计算机可读存储介质
CN113891040A (zh) * 2021-09-24 2022-01-04 深圳Tcl新技术有限公司 视频处理方法、装置、计算机设备和存储介质
CN113938698A (zh) * 2021-10-19 2022-01-14 广州方硅信息技术有限公司 直播用户资料的显示控制方法、装置及计算机设备
CN115086759A (zh) * 2022-05-13 2022-09-20 北京达佳互联信息技术有限公司 视频处理方法、装置、计算机设备及介质
CN115086774A (zh) * 2022-05-31 2022-09-20 北京达佳互联信息技术有限公司 资源显示方法、装置、电子设备及存储介质
WO2023169049A1 (fr) * 2022-03-09 2023-09-14 聚好看科技股份有限公司 Dispositif d'affichage et serveur

Families Citing this family (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109034115B (zh) * 2018-08-22 2021-10-22 Oppo广东移动通信有限公司 视频识图方法、装置、终端及存储介质
CN109857309B (zh) * 2019-01-21 2022-02-01 武汉卡比特信息有限公司 一种基于iOS移动终端的屏幕截图投射方法
CN110134807B (zh) * 2019-05-17 2021-06-04 苏州科达科技股份有限公司 目标检索方法、装置、系统及存储介质
CN110442806B (zh) * 2019-08-05 2022-04-26 百度在线网络技术(北京)有限公司 用于识别图像的方法和装置
CN112784137A (zh) * 2019-11-04 2021-05-11 海信视像科技股份有限公司 显示设备、显示方法及计算设备
CN110909776A (zh) * 2019-11-11 2020-03-24 维沃移动通信有限公司 一种图像识别方法及电子设备
CN111339395A (zh) * 2020-02-11 2020-06-26 山东经贸职业学院 一种用于电子商务系统的数据信息匹配方法和系统
CN111652678B (zh) * 2020-05-27 2023-11-14 腾讯科技(深圳)有限公司 物品信息显示方法、装置、终端、服务器及可读存储介质
CN112162672A (zh) * 2020-10-19 2021-01-01 腾讯科技(深圳)有限公司 信息流的显示处理方法、装置、电子设备及存储介质
CN112996196B (zh) * 2021-02-04 2023-02-10 沃特威(广州)电子科技有限公司 智能环境灯光控制方法、系统、计算机设备及存储介质
CN113282769A (zh) * 2021-04-25 2021-08-20 维沃移动通信有限公司 多媒体文件的处理方法、装置及电子设备
CN113282768A (zh) * 2021-04-25 2021-08-20 维沃移动通信有限公司 多媒体文件处理方法、装置及电子设备
CN113110785B (zh) * 2021-05-12 2023-04-18 北京字节跳动网络技术有限公司 图像处理方法、装置、电子设备以及计算机可读存储介质
CN113761360A (zh) * 2021-05-27 2021-12-07 腾讯科技(深圳)有限公司 基于视频的物品搜索方法、装置、设备及存储介质
CN115527135A (zh) * 2021-06-24 2022-12-27 Oppo广东移动通信有限公司 内容识别方法、装置以及电子设备
CN115878838A (zh) * 2021-09-27 2023-03-31 北京有竹居网络技术有限公司 基于视频的信息展示方法、装置、电子设备及存储介质
CN114268847A (zh) * 2021-12-15 2022-04-01 北京百度网讯科技有限公司 视频播放方法、装置、电子设备及存储介质
CN116431947A (zh) * 2022-01-04 2023-07-14 腾讯科技(深圳)有限公司 多媒体处理方法、装置、设备、介质及计算机程序产品

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106028160A (zh) * 2016-06-03 2016-10-12 腾讯科技(深圳)有限公司 一种图像数据处理方法及其设备
CN106202316A (zh) * 2016-07-01 2016-12-07 传线网络科技(上海)有限公司 基于视频的商品信息获取方法及装置
US20160357406A1 (en) * 2015-06-05 2016-12-08 Samsung Electronics Co., Ltd. Operating method for image and electronic device supporting the same
CN107105340A (zh) * 2017-03-21 2017-08-29 百度在线网络技术(北京)有限公司 基于人工智能的视频中显示人物信息方法、装置和系统
CN107515868A (zh) * 2016-06-15 2017-12-26 北京陌上花科技有限公司 搜索方法及装置
CN107957891A (zh) * 2017-11-22 2018-04-24 暴风集团股份有限公司 一种视频播放器信息显示方法、装置、终端及系统
CN108089786A (zh) * 2017-12-14 2018-05-29 广东欧珀移动通信有限公司 用户界面显示方法、装置、设备及存储介质
CN109034115A (zh) * 2018-08-22 2018-12-18 Oppo广东移动通信有限公司 视频识图方法、装置、终端及存储介质

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8977639B2 (en) * 2009-12-02 2015-03-10 Google Inc. Actionable search results for visual queries
US9852156B2 (en) * 2009-12-03 2017-12-26 Google Inc. Hybrid use of location sensor data and visual query to return local listings for visual query
JP2013200793A (ja) * 2012-03-26 2013-10-03 Sony Corp 情報処理装置、情報処理方法およびプログラム
CN102682091A (zh) * 2012-04-25 2012-09-19 腾讯科技(深圳)有限公司 基于云服务的视觉搜索方法和系统
US20150089446A1 (en) * 2013-09-24 2015-03-26 Google Inc. Providing control points in images
KR102158691B1 (ko) * 2014-01-08 2020-09-22 엘지전자 주식회사 이동 단말기 및 그 제어 방법
CN104090762B (zh) * 2014-07-10 2017-04-19 福州瑞芯微电子股份有限公司 一种截图处理装置和方法
US10664515B2 (en) * 2015-05-29 2020-05-26 Microsoft Technology Licensing, Llc Task-focused search by image
CN106529413A (zh) * 2016-10-13 2017-03-22 北京小米移动软件有限公司 信息获取方法及装置
CN107256109B (zh) * 2017-05-27 2021-03-16 北京小米移动软件有限公司 信息显示方法、装置及终端

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160357406A1 (en) * 2015-06-05 2016-12-08 Samsung Electronics Co., Ltd. Operating method for image and electronic device supporting the same
CN106028160A (zh) * 2016-06-03 2016-10-12 腾讯科技(深圳)有限公司 一种图像数据处理方法及其设备
CN107515868A (zh) * 2016-06-15 2017-12-26 北京陌上花科技有限公司 搜索方法及装置
CN106202316A (zh) * 2016-07-01 2016-12-07 传线网络科技(上海)有限公司 基于视频的商品信息获取方法及装置
CN107105340A (zh) * 2017-03-21 2017-08-29 百度在线网络技术(北京)有限公司 基于人工智能的视频中显示人物信息方法、装置和系统
CN107957891A (zh) * 2017-11-22 2018-04-24 暴风集团股份有限公司 一种视频播放器信息显示方法、装置、终端及系统
CN108089786A (zh) * 2017-12-14 2018-05-29 广东欧珀移动通信有限公司 用户界面显示方法、装置、设备及存储介质
CN109034115A (zh) * 2018-08-22 2018-12-18 Oppo广东移动通信有限公司 视频识图方法、装置、终端及存储介质

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111444822A (zh) * 2020-03-24 2020-07-24 北京奇艺世纪科技有限公司 对象识别方法和装置、存储介质和电子装置
CN111444822B (zh) * 2020-03-24 2024-02-06 北京奇艺世纪科技有限公司 对象识别方法和装置、存储介质和电子装置
CN111541907B (zh) * 2020-04-23 2023-09-22 腾讯科技(深圳)有限公司 物品显示方法、装置、设备及存储介质
CN111541907A (zh) * 2020-04-23 2020-08-14 腾讯科技(深圳)有限公司 物品显示方法、装置、设备及存储介质
CN112565863A (zh) * 2020-11-26 2021-03-26 深圳Tcl新技术有限公司 视频播放方法、装置、终端设备及计算机可读存储介质
CN112584213A (zh) * 2020-12-11 2021-03-30 海信视像科技股份有限公司 一种显示设备和图像识别结果的展示方法
CN113747182A (zh) * 2021-01-18 2021-12-03 北京京东拓先科技有限公司 物品展示方法、客户端、直播服务器和计算机存储介质
CN112801004A (zh) * 2021-02-05 2021-05-14 网易(杭州)网络有限公司 视频片段的筛选方法、装置、设备及存储介质
CN113766297B (zh) * 2021-05-27 2023-12-05 腾讯科技(深圳)有限公司 视频处理方法、播放终端及计算机可读存储介质
CN113766297A (zh) * 2021-05-27 2021-12-07 腾讯科技(深圳)有限公司 视频处理方法、播放终端及计算机可读存储介质
CN113891040A (zh) * 2021-09-24 2022-01-04 深圳Tcl新技术有限公司 视频处理方法、装置、计算机设备和存储介质
CN113938698A (zh) * 2021-10-19 2022-01-14 广州方硅信息技术有限公司 直播用户资料的显示控制方法、装置及计算机设备
CN113938698B (zh) * 2021-10-19 2024-03-12 广州方硅信息技术有限公司 直播用户资料的显示控制方法、装置及计算机设备
WO2023169049A1 (fr) * 2022-03-09 2023-09-14 聚好看科技股份有限公司 Dispositif d'affichage et serveur
CN115086759A (zh) * 2022-05-13 2022-09-20 北京达佳互联信息技术有限公司 视频处理方法、装置、计算机设备及介质
CN115086774A (zh) * 2022-05-31 2022-09-20 北京达佳互联信息技术有限公司 资源显示方法、装置、电子设备及存储介质
CN115086774B (zh) * 2022-05-31 2024-03-05 北京达佳互联信息技术有限公司 资源显示方法、装置、电子设备及存储介质

Also Published As

Publication number Publication date
CN109034115A (zh) 2018-12-18
CN109034115B (zh) 2021-10-22

Similar Documents

Publication Publication Date Title
WO2020038167A1 (fr) Procédé et appareil de reconnaissance d'image vidéo, terminal et support d'enregistrement
US11320960B2 (en) Icon display method, device, and terminal
US11467715B2 (en) User interface display method, terminal and non-transitory computer-readable storage medium for splitting a display using a multi-finger swipe
US11412153B2 (en) Model-based method for capturing images, terminal, and storage medium
US11644943B2 (en) Method for icon display, terminal, and storage medium
WO2020038168A1 (fr) Procédé et dispositif de partage de contenu, terminal et support de stockage
WO2019174477A1 (fr) Procédé et dispositif d'affichage d'interface utilisateur, et terminal
US11500513B2 (en) Method for icon display, terminal, and storage medium
WO2020156199A1 (fr) Procédé et dispositif de connexion à une application, terminal et support de stockage
WO2019091411A1 (fr) Procédé de capture d'image, dispositif, terminal et support de stockage
CN109992315B (zh) 触摸屏控制方法、装置、终端及存储介质
WO2019233307A1 (fr) Procédé et appareil d'affichage d'interface utilisateur, terminal et support d'informations
WO2019047738A1 (fr) Procédé d'affichage de messages, dispositif, terminal mobile et support de stockage
CN111767554B (zh) 屏幕分享方法、装置、存储介质及电子设备
WO2022001452A1 (fr) Procédé et appareil d'affichage d'informations, dispositif à porter sur soi, et support de stockage
CN113268212A (zh) 投屏方法、装置、存储介质及电子设备
CN112035046B (zh) 榜单信息显示方法、装置、电子设备及存储介质
WO2019047183A1 (fr) Procédé d'affichage de touche, appareil et terminal
US20200244869A1 (en) Method for capturing images, terminal, and storage medium
CN109683760B (zh) 最近内容的显示方法、装置、终端及存储介质
CN111127469A (zh) 缩略图显示方法、装置、存储介质以及终端
CN111526290B (zh) 图像处理方法、装置、终端及存储介质
WO2021254113A1 (fr) Procédé de commande destiné à une interface tridimensionnelle et terminal
US11194598B2 (en) Information display method, terminal and storage medium
CN110971974B (zh) 配置参数创建方法、装置、终端及存储介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19852274

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19852274

Country of ref document: EP

Kind code of ref document: A1