CN117294900A - Video playing method and device, electronic equipment and readable storage medium - Google Patents

Video playing method and device, electronic equipment and readable storage medium Download PDF

Info

Publication number
CN117294900A
CN117294900A CN202311385736.0A CN202311385736A CN117294900A CN 117294900 A CN117294900 A CN 117294900A CN 202311385736 A CN202311385736 A CN 202311385736A CN 117294900 A CN117294900 A CN 117294900A
Authority
CN
China
Prior art keywords
video
identifier
input
location
display
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202311385736.0A
Other languages
Chinese (zh)
Inventor
陈艳玲
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Vivo Mobile Communication Co Ltd
Original Assignee
Vivo Mobile Communication Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Vivo Mobile Communication Co Ltd filed Critical Vivo Mobile Communication Co Ltd
Priority to CN202311385736.0A priority Critical patent/CN117294900A/en
Publication of CN117294900A publication Critical patent/CN117294900A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47217End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for controlling playback functions for recorded or on-demand content, e.g. using progress bars, mode or play-point indicators or bookmarks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/485End-user interface for client configuration
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8456Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments

Abstract

The application discloses a video playing method, a video playing device, electronic equipment and a readable storage medium, which belong to the technical field of electronics, and the method comprises the following steps: receiving a first input of a user to a first place identifier in at least one place identifier under the condition that the at least one place identifier corresponding to the first video is displayed; responsive to a first input, playing a first video clip corresponding to a first location identifier; wherein the first video comprises at least two video clips, one location identifier corresponds to the at least one video clip, and one location identifier indicates one location in the first video.

Description

Video playing method and device, electronic equipment and readable storage medium
Technical Field
The application belongs to the technical field of electronics, and particularly relates to a video playing method, a video playing device, electronic equipment and a readable storage medium.
Background
Currently, when a user views a video, the user can only drag a video progress bar to position a video clip, so as to select the video content to be watched or skip the video content not to be watched.
However, when the user drags the video progress bar to locate the video clip, the user needs to drag the video progress bar first to view the corresponding video content, so as to determine whether the corresponding video clip is the video clip that wants to be viewed. If the corresponding video clip after dragging is not the video content to be watched, the user needs to drag the video progress bar again to find the video content to be watched. Thus, the video progress bar needs to be dragged repeatedly, and the wanted video clip cannot be positioned quickly.
Disclosure of Invention
An object of the embodiments of the present application is to provide a video playing method, apparatus, electronic device, and readable storage medium, which can quickly locate a video clip to be watched.
In a first aspect, an embodiment of the present application provides a video playing method, where the method includes: receiving a first input of a user to a first place identifier in at least one place identifier under the condition that the at least one place identifier corresponding to the first video is displayed; responsive to a first input, playing a first video clip corresponding to a first location identifier; wherein the first video comprises at least two video clips, one location identifier corresponds to the at least one video clip, and one location identifier indicates one location in the first video.
In a second aspect, an embodiment of the present application provides a video playing device, including: the device comprises a display module, a receiving module and a processing module; the receiving module is used for receiving a first input of a user to a first place identifier in the at least one place identifier under the condition that the display module displays the at least one place identifier corresponding to the first video; the processing module is used for responding to the first input received by the receiving module and playing a first video clip corresponding to the first place identifier; wherein the first video comprises at least two video clips, one location identifier corresponds to the at least one video clip, and one location identifier indicates one location in the first video.
In a third aspect, embodiments of the present application provide an electronic device comprising a processor and a memory storing a program or instructions executable on the processor, which when executed by the processor, implement the steps of the method as described in the first aspect.
In a fourth aspect, embodiments of the present application provide a readable storage medium having stored thereon a program or instructions which when executed by a processor implement the steps of the method according to the first aspect.
In a fifth aspect, embodiments of the present application provide a chip, where the chip includes a processor and a communication interface, where the communication interface is coupled to the processor, and where the processor is configured to execute a program or instructions to implement a method according to the first aspect.
In a sixth aspect, embodiments of the present application provide a computer program product stored in a storage medium, the program product being executable by at least one processor to implement the method according to the first aspect.
In the embodiment of the application, the location identifier corresponding to the location in the video can be displayed, the user selects the location identifier of interest to play the video clip corresponding to the selected location identifier, and compared with the video clip which the user needs to drag the video progress bar repeatedly to search for the video clip which the user wants to watch in the related technology, the operation process of the embodiment of the application is simple, and the video clip which the user wants to watch can be positioned fast.
Drawings
Fig. 1 is one of flowcharts of a video playing method provided in an embodiment of the present application;
FIG. 2 is a second flowchart of a video playing method according to the embodiment of the present application;
FIG. 3 is an interface schematic diagram of a video map according to an embodiment of the present disclosure;
FIG. 4 (A) is a schematic diagram of an interface for displaying a menu list according to an embodiment of the present application;
FIG. 4 (B) is a schematic diagram of an interface for displaying a video map according to an embodiment of the present application;
FIG. 5 is a schematic interface diagram of an edit location identifier provided in an embodiment of the present application;
FIG. 6 is a schematic diagram of an interface for magnifying and displaying a video map according to an embodiment of the present application;
FIG. 7 is a third flowchart of a video playing method according to the embodiment of the present application;
fig. 8 is an interface schematic diagram of an electronic device displaying a playing route according to an embodiment of the present application;
FIG. 9 is a fourth flowchart of a video playing method according to the embodiment of the present application;
fig. 10 is an interface schematic diagram of an electronic device displaying object identifiers and scenario identifiers according to an embodiment of the present application;
fig. 11 is a schematic structural diagram of a video playing device according to an embodiment of the present application;
fig. 12 is a schematic diagram of a hardware structure of an electronic device according to an embodiment of the present application;
Fig. 13 is a second schematic diagram of a hardware structure of an electronic device according to an embodiment of the present application.
Detailed Description
Technical solutions in the embodiments of the present application will be clearly described below with reference to the drawings in the embodiments of the present application, and it is apparent that the described embodiments are some embodiments of the present application, but not all embodiments. All other embodiments obtained by a person of ordinary skill in the art based on the embodiments in the present application are within the scope of the protection of the present application.
The terms "first," "second," and the like in the description of the present application, are used for distinguishing between similar objects and not necessarily for describing a particular sequential or chronological order. It is to be understood that the terms so used are interchangeable under appropriate circumstances such that the embodiments of the application are capable of operation in sequences other than those illustrated or otherwise described herein, and that the objects identified by "first," "second," etc. are generally of a type and do not limit the number of objects, for example, the first object may be one or more. In addition, "and/or" in the specification means at least one of the connected objects, and the character "/", generally means a relationship in which the associated objects are one kind of "or".
The terms "at least one", and the like in the description of the present application refer to any one, any two, or a combination of two or more of the objects that it comprises. For example, at least one of a, b, c (item) may represent: "a", "b", "c", "a and b", "a and c", "b and c" and "a, b and c", wherein a, b, c may be single or plural. Similarly, the term "at least two" means two or more, and the meaning of the expression is similar to the term "at least one".
The video playing method, the video playing device, the electronic equipment and the readable storage medium provided by the embodiment of the application are described in detail below through specific embodiments and application scenes thereof with reference to the accompanying drawings.
The video playing method, the video playing device, the electronic equipment and the readable storage medium can be applied to scenes for playing videos.
With the development of communication technology, video is a good information propagation carrier because of synchronous transmission of audio, video and text, so that video application programs for playing video are increasingly widely applied and take up the mainstream position in social networks.
However, when watching video, the user can only play and select video clips according to the number of video sets or dragging the video progress bar. For example, in the archaeological documentaries, the video content of the recall or the explanation of the staff is often inserted in the archaeological video clip, and the user needs to drag the video progress bar repeatedly to select the video content only including the archaeology. Thus, the video progress bar needs to be dragged repeatedly, and the wanted video clip cannot be positioned quickly.
In the video playing method, the device, the electronic equipment and the readable storage medium provided by the embodiment of the application, the place identifier corresponding to the place in the video can be displayed, the user selects the place identifier of interest to play the video clip corresponding to the selected place identifier, and compared with the video clip which is required to be repeatedly dragged by the user to search for the video clip which is required to be watched by the user in the related technology, the operation process of the embodiment of the application is simple, and the video clip which is required to be watched by the user can be rapidly positioned.
The execution subject of the video playing method provided in the embodiment of the present application may be a video playing device. The video playing device may be an electronic device, or may be a component in the electronic device, such as an integrated circuit or a chip, for example. The video playing method provided in the embodiment of the present application will be exemplarily described below by taking an electronic device as an example.
An embodiment of the present application provides a video playing method, and fig. 1 shows a flowchart of the video playing method provided in the embodiment of the present application, where the method may be applied to an electronic device. As shown in fig. 1, the video playing method provided in the embodiment of the present application may include the following steps 101 and 102.
Step 101, the electronic device receives a first input of a user to a first place identifier in at least one place identifier under the condition that at least one place identifier corresponding to a first video is displayed.
Step 102, the electronic device responds to the first input to play the first video clip corresponding to the first place identifier.
Wherein the first video may include at least two video clips, one location identifier corresponds to the at least one video clip, and one location identifier indicates one location in the first video.
In some embodiments of the present application, the first video may be a video obtained by the electronic device from another device, a video stored in the electronic device, or a video captured by the electronic device.
It can be appreciated that playing the video in the video application program by the electronic device may actually play the online video acquired from the server for the electronic device, or may be the video stored in the electronic device.
The first video may be, for example, any movie video in a video application.
The first video may be, for example, a documentary video stored by the electronic device.
In some embodiments of the present application, the video clips may be video clips that include scenario of different locations in the first video.
Illustratively, the first video is taken as an archaeological recording sheet. The archaeological documentaries can comprise video clips of archaeological places of staff and video clips of staff's commentary in a studio.
Illustratively, taking the first video as a movie. The movie may be divided into a plurality of different video clips according to the location where the scenario occurs.
In some embodiments of the present application, the location identifier may be an identifier of a location in the first video where the scenario occurs. That is, the location identification may be an identification of a location associated with the video clip.
For example, the location identifier may be an identifier of a location where a scenario occurs.
In some embodiments of the present application, the first location identifier may be one or more of at least one location identifier corresponding to the first video.
It can be appreciated that the electronic device can determine, through the first input of the user to the first location identifier, that the video clip that the user wants to view is a video clip associated with the location corresponding to the first location identifier, so that the video clip can be accurately selected, and the first video clip corresponding to the first location identifier can be directly played. And the user does not need to drag the video progress bar repeatedly to search the video clips to be watched, so that the positioning efficiency of the video clips is improved.
In some embodiments of the present application, the first input may be used to trigger the electronic device to select a location identifier corresponding to a video clip to be played. In other words, the first input may be a selection input of a location identification.
In some embodiments of the present application, the first input includes, but is not limited to: the user inputs the touch control of the electronic device screen through the touch control device such as a finger or a stylus pen, or inputs a voice command input by the user, or inputs a specific gesture input by the user, or inputs other feasibility, and the specific touch control input can be determined according to actual use requirements. The specific gesture in the embodiment of the present application may be any one of a single-click gesture, a sliding gesture, a dragging gesture, a pressure recognition gesture, a long-press gesture, an area change gesture, a double-press gesture, and a double-click gesture; the click input in the embodiment of the application may be single click input, double click input, or any number of click inputs, and may also be long press input or short press input.
For example, the electronic device may receive a click input of the first location identifier from the user, and play the first video clip corresponding to the first location identifier.
For example, the electronic device may receive a drag input of a user to a first location identifier, drag the first location identifier to a playing interface of the first video, and play a first video clip corresponding to the first location identifier.
In some embodiments of the present application, the first location identifier may include at least two location identifiers.
In the video playing method, the location identifier corresponding to the location in the video can be displayed, the user selects the location identifier of interest to play the video clip corresponding to the selected location identifier, and compared with the video clip which needs to be repeatedly dragged by the user to search for the video clip which needs to be watched by the user in the related technology, the operation process of the video playing method is simple, and the video clip which needs to be watched by the user can be rapidly positioned.
In some embodiments of the present application, as shown in fig. 2 in conjunction with fig. 1, before the step 101, the video playing method provided in the embodiments of the present application may further include the following steps 103 and 104.
Step 103, the electronic device receives a second input from the user.
Step 104, the electronic device responds to the second input to display a video map corresponding to the first video.
Wherein the video map may include at least one location identifier.
In some embodiments of the present application, the video map may be a map including location identifiers corresponding to at least two video segments in the first video. That is, the user can quickly know the place appearing in the first video through the video map.
In some embodiments of the present application, the relative display position between at least one location identifier in the video map may be determined by the geographic position between locations, may be set by a user, or may be determined by a default display mode of the electronic device. The embodiments of the present application are not particularly limited.
In some embodiments of the present application, the "geographic location between places" may be a geographic location where the place corresponding to the place is located in the first video for the place identification.
Illustratively, the sites in the first video include site a, site b, and site c. If, in the first video, location a is located north-east of location b and location c is located west of location b, the electronic device may display the video map 30 as shown in fig. 3. In the video map 30, the location mark 31 of the location b is located at the center position of the video map 30, the location mark 32 of the location a is located in the northeast of the location mark 31, and the location mark 33 of the location c is located in the west of the location mark 31.
In some embodiments of the present application, in a case where the geographic location of any location identifier in the video map cannot be obtained, the electronic device may determine, according to the frequency of occurrence of the location identifier corresponding to the location identifier in the first video, a display position of the location identifier, where the higher the frequency of occurrence, the closer the display position is to the center of the video map, and the lower the frequency of occurrence, the location identifier corresponding to the location identifier may be displayed at an edge position of the video map.
In some embodiments of the present application, the second input may be used to trigger the electronic device to display a video map corresponding to the first video. In other words, the second input may be a display input of a video map corresponding to the first video.
In some embodiments of the present application, the second input includes, but is not limited to: the user inputs the touch control of the electronic device screen through the touch control device such as a finger or a stylus pen, or inputs a voice command input by the user, or inputs a specific gesture input by the user, or inputs other feasibility, and the specific touch control input can be determined according to actual use requirements. The specific gesture in the embodiment of the present application may be any one of a single-click gesture, a sliding gesture, a dragging gesture, a pressure recognition gesture, a long-press gesture, an area change gesture, a double-press gesture, and a double-click gesture; the click input in the embodiment of the application may be single click input, double click input, or any number of click inputs, and may also be long press input or short press input.
For example, the electronic device may receive a long press input from a user to a playing interface of the first video, and display a video map corresponding to the first video.
For example, the electronic device can display a menu list control at a play interface of the first video, and the electronic device can receive user input to the menu list to display a video map including at least one location identifier.
For example, as shown in fig. 4 (a), in the case of playing the video 40, the electronic device may receive user input to the menu list control 41, displaying the menu list 42. Then, as shown in fig. 4 (B), the electronic device may receive user input to the video map in the menu list 42, and display the video map 43 containing 9 place identifications of the place identification a, the place identification B, the place identification C, the place identification D, the place identification E, the place identification F, the place identification G, the place identification J, and the place identification K.
In some embodiments of the present application, the electronic device may acquire and display the video map corresponding to the first video through the following two embodiments.
A first possible implementation: and automatically generating a video map corresponding to the first video.
In some embodiments of the present application, the electronic device may determine location information associated with each video clip based on video content of the video clip in the first video. The electronic device may then generate a corresponding location identification based on the location information, thereby generating and displaying a video map corresponding to the first video.
In some embodiments of the present application, the video content of the video clip may include at least one of: subtitle information in the video clip, and scene information of a video scene corresponding to the video clip.
In some embodiments of the present application, the location information may include a location name or scene information.
In some embodiments of the present application, the electronic device may correspond all video clips in the first video to the location identifier according to the location name and the scene information.
In some embodiments of the present application, the subtitle information in the video clip may be non-visual content displayed in text form in the video clip, such as text of a person dialogue, a bystanding, an annotation, and the like in the video clip.
In some embodiments of the present application, the electronic device may analyze the subtitle information for the video clip through artificial intelligence (Artificial Intelligence, AI) to obtain a place name associated with the video clip.
It is understood that the electronic device may display the location identifier corresponding to the location after obtaining the location through the caption information.
In some embodiments of the present application, the scene information of the video scene corresponding to the video clip may include scene arrangements of different places in the video clip, such as indoor table and chair placement mode, outdoor public facilities, and the like.
It will be appreciated that typically identical video scenes will correspond to identical locations in a video clip. Therefore, the electronic device can match scene information of a plurality of video scenes and determine video clips belonging to the same video scene, so that the video clips corresponding to the same location can be determined.
It should be noted that, for a video scene where a place name cannot be obtained through subtitle information, the electronic device may name the place with a default place name, and correspond to a video clip associated with the place with the default place name.
In some embodiments of the present application, after generating the location identifier corresponding to the location, the electronic device may display a video map including the location identifier, so that the user may select a video clip corresponding to the location identifier in the video map.
In this way, since the electronic device can determine the location information through the subtitle information in the video clip and the scene information of the video scene, the corresponding location identifier is generated. Therefore, the electronic equipment can respectively correspond all video clips in the video to the associated location identifiers, so that a user can accurately select the video clips through selecting the location identifiers, and the positioning efficiency of the video clips is improved.
Another embodiment: and acquiring a video map corresponding to the first video from other equipment.
In some embodiments of the present application, the electronic device may directly display the video map generated by the other device. In other words, the video map displayed by the electronic device may be generated for other devices.
For example, after the video map corresponding to the first video is generated, the other devices may directly send the video map to the electronic device of the user, so that the electronic device of the user may directly display the video map without regenerating the video map.
For example, after generating the video map corresponding to the first video, the other devices may upload the video map to the server, and the electronic device may download and display the video map corresponding to the first video from the server.
Therefore, the electronic equipment can directly acquire the video map generated by other equipment for display, so that resources and time required by the electronic equipment for generating the video map can be saved.
In some embodiments of the present application, the electronic device may display the video map in any display area of the screen.
For example, the electronic device may display the video map in an upper left corner of the playback interface of the first video.
For example, the electronic device may display the video map in a display area other than the playback interface of the first video.
For example, if the electronic device includes at least two screens, the electronic device may display the playback interface of the first video on one screen and the video map on the other screen.
In some embodiments of the present application, the electronic device may display the video map in any display form.
For example, the electronic device may display the video map in the form of a floating window.
The electronic device may also display the video map directly superimposed on the display interface of the first video.
Therefore, the electronic equipment can display the video map containing the location mark, so that a user can quickly know the location of the scenario in the video through the video map, and further can quickly and accurately select the video clip to be watched through the location of the scenario in the video, and the positioning efficiency of the video clip is improved.
In some embodiments of the present application, after the step 104, the video playing method provided in the embodiments of the present application may further include the following steps 105 and 106.
Step 105, the electronic device receives a third input from the user.
Step 106, the electronic device responds to the third input to update the display parameters of the video map.
Wherein, the display parameters may include at least one of the following: display position, display size, display shape, display color, display position of at least one location indicator, name of at least one location indicator.
In some embodiments of the present application, the display position, display size, display shape, and display color of the video map may be display parameters of the whole video map, or may be display parameters of a location identifier in the video map. The embodiment of the present application is not particularly limited.
For example, after the electronic device displays the video map, the user may change the display position of the video map through the third input; the display position of the location identifier in the video map may also be changed by the third input.
For example, after the electronic device displays the video map, the user may adjust the display size of the video map through the third input; the display size of the location identifier in the video map may also be adjusted via the third input.
In some embodiments of the present application, after the electronic device displays the video map, the user may edit the display parameters associated with the at least one location identifier in the video map via a third input to display the updated at least one location identifier.
In some embodiments of the present application, the user may update all display parameters of the video map or update display parameters of a partial display area of the video map through the third input. It will be appreciated that the electronic device updates the display parameters of the video map, in effect the display parameters associated with the location identification in the video map.
It will be appreciated that there may be situations where the location identifier corresponds inaccurately to the video clip in the at least one location identifier displayed by the electronic device, or where the shape, color, etc. of the at least one location identifier displayed by the electronic device is disliked by the user. Thus, the user may edit the at least one location identifier via the third input to obtain a final at least one location identifier.
In some embodiments of the present application, the third input may be an input editing a display parameter associated with a location identifier in the video map. In other words, the third input may be an edit input of a location identifier associated with a display parameter in the video map.
In some embodiments of the present application, upon updating the display parameters associated with the location identifier in the video map, the electronic device may, after receiving the third input from the user, adjust a display position, a display size, a display shape, a display color, a display position of the at least one location identifier, or a name of the at least one location identifier associated with the location identifier in the video map.
For example, the electronic device may receive input from a user dragging one location identifier to another location identifier, display a display location of the one location identifier on the other location identifier, and merge the two video clips.
Illustratively, as shown in FIG. 5, the electronic device may receive a third input by the user of the location identifier for location A in the video map, editing the display shape of the location identifier for location A into shape 50.
Illustratively, the electronic device may also receive a third input by the user of location identifier B in the video map, changing the display color of location identifier B. Such as setting the display color of the place mark B to dark green.
Illustratively, the electronic device may also receive a third input by the user of the location identifier C in the video map, editing the name of the location identifier C. The name of the location mark C is edited from "X temple" to "Y temple".
In some embodiments of the present application, the third input includes, but is not limited to: the user inputs the touch control of the electronic device screen through the touch control device such as a finger or a stylus pen, or inputs a voice command input by the user, or inputs a specific gesture input by the user, or inputs other feasibility, and the specific touch control input can be determined according to actual use requirements. The specific gesture in the embodiment of the present application may be any one of a single-click gesture, a sliding gesture, a dragging gesture, a pressure recognition gesture, a long-press gesture, an area change gesture, a double-press gesture, and a double-click gesture; the click input in the embodiment of the application may be single click input, double click input, or any number of click inputs, and may also be long press input or short press input.
For example, the electronic device may receive input that the user double clicks on the location identifier in the video map, changing the display shape of the location identifier.
For example, the electronic device may receive input from a user pressing a location identifier in a video map for a long time, changing the display color of the location identifier.
In some embodiments of the present application, when changing the name of the location identifier in the video map, if the changed name is the same as an existing name in the video map, the electronic device may combine the video segment corresponding to the location identifier with the video segment corresponding to the location identifier corresponding to the existing name in the video map.
Illustratively, take the video map containing a location identifier 1 and a location identifier 2, where location identifier 1 corresponds to video segment 1 and location identifier 2 corresponds to video segment 2 as an example. If the user modifies the name of location identifier 1 to the name of location identifier 2, the electronic device may merge video clip 1 with video clip 2 and cancel the display of location identifier 1. That is, the location identifier 2 corresponds to the video clip 1 and the video clip 2.
In some embodiments of the present application, when the electronic device updates the display size of the video map, the electronic device may update at least one of an adjustment direction of the video map and a zoom level of the video map.
In some embodiments of the present application, the direction of adjustment of the video map may indicate the direction in which the video map zooms.
Illustratively, the adjustment direction of the video map is taken as a left enlargement. The electronic device may enlarge the display size of the left area of the video map.
Illustratively, the adjustment direction of the video map is taken as a downward zoom-out example. The electronic device may reduce the display size of the lower region of the video map.
In some embodiments of the present application, the degree of zoom of the video map may indicate the scale at which the video map is zoomed out and zoomed in.
In some embodiments of the present application, the electronic device may determine an adjustment direction of the video map according to the input direction of the third input; and determining the zoom degree of the video map according to the input duration of the third input.
The third input may be, for example, a user's press input to a video map displayed by the electronic device to trigger the electronic device to zoom in on the map content of the video map at the center of the pressed location.
For example, the third input may be a two-finger pinch input of a video map displayed by the electronic device by the user to trigger the electronic device to zoom out the map content of the video map toward the center of the two-finger pinch.
In some embodiments of the present application, the map content may include location identifiers in the video map and relative display positions between the location identifiers.
In some embodiments of the present application, the electronic device may adjust and display map content within the target area in the video map according to the adjustment direction and the zoom level determined according to the third input.
In some embodiments of the present application, the target area may be an area to be adjusted selected by the user through three inputs in the video map.
Illustratively, the third input is exemplified as a user's press input to a video map displayed by the electronic device. The target area may be an area corresponding to the pressed position in the video map.
Illustratively, the third input is exemplified as a two-finger pinch input of a video map displayed by the user to the electronic device. The target area may be an area between two fingers in the video map.
Illustratively, assume a scale 1 for the video map: and 5, the electronic equipment can receive a third input of the user on the video map, determine that the adjustment direction of the video map is leftward, and the zoom degree is 1:10. the electronic device can set the target area in the video map corresponding to the third input to 1:10.
For example, assuming that the video map contains location identifications corresponding to all locations in the first video, the electronic device may receive a third input from the user to the video map, determine an adjustment direction and a degree of scaling of the video map. Then, the electronic device may display the location identifier corresponding to the partial location in the first video according to the adjustment direction and the zoom level. That is, the electronic device adjusts the globally displayed video map to the locally displayed video map.
The adjustment of the display size described above is an adjustment of the scale of the map content displayed on the target area in the video map, and is not an adjustment of the display size of the video map.
In practical implementation, the electronic device may also receive user input of the video map, and adjust the display size of the video map.
For example, as shown in fig. 6, the electronic device may receive a click input of the user on the video map 61, and display the video map 61 in an enlarged manner, so as to obtain the video map 62, so that the user may view the map identifier in the video map more clearly.
The steps 105 and 106 may be performed before the steps 101 and 102, after the steps 101 and 102, or simultaneously with the steps 101 and 102. The embodiments of the present application are not particularly limited.
Therefore, the electronic equipment can update the display information in the video map according to the input of the user, so that the flexibility and the diversity of the place identification in the video map displayed by the electronic equipment can be improved.
In some embodiments of the present application, as shown in fig. 7 in conjunction with fig. 1, after the step 101, the video playing method provided in the embodiments of the present application may further include the following step 107.
Step 107, the electronic device responds to the first input to update the display parameter of the first playing guidance identifier corresponding to the first location identifier.
In some embodiments of the present application, the video map may include at least two play guide identifiers, where the play guide identifiers may be used to indicate at least two video clips and a play order of the at least two video clips. In other words, the video map may include at least one location identifier, and an identifier of a play order of the video clip corresponding to the at least one location identifier.
In some embodiments of the present application, the playing order of the video clips may be a playing order of the video clips arranged according to a time sequence of a location change. In other words, by playing the guide identifier, the user can learn the play route of the place change in the first video.
Illustratively, the first video includes a video segment a corresponding to the location identifier a, a video segment B corresponding to the location identifier B, and a video segment C corresponding to the location identifier C. As shown in fig. 8, the electronic device may indicate that the next video clip of the video clip a is the video clip C through the play guidance mark 81, and indicate that the next video clip of the video clip C is the video clip B through the play guidance mark 82. That is, in the first video, the place following the place corresponding to the place identifier a is the place corresponding to the place identifier C, and the place following the place corresponding to the place identifier C is the place corresponding to the place identifier B.
In some embodiments of the present application, the first play guide identifier may be one or more of at least two play guide identifiers.
In some embodiments of the present application, the "first play guiding identifier corresponding to the first location identifier" may be understood as: the at least two video clips indicated by the first playing guide identifier comprise video clips corresponding to the first place identifier. That is, the first play guide identifier may indicate a play order of the video clip corresponding to the first location identifier and other video clips.
In some embodiments of the present application, the display parameter of the play guidance identifier may indicate a play status of the video clip indicated by the play guidance identifier.
For example, the electronic device may set the play guide identifier corresponding to the video clip in the play state to be highlighted.
For example, the electronic device may set the play guidance identifier corresponding to the video clip that is not in the play state to be displayed in gray.
In some embodiments of the present application, after receiving the first input of the user, the electronic device may update, when playing the first video clip corresponding to the first location identifier, a display parameter of the first play guidance identifier corresponding to the first location identifier, and display the first play guidance identifier differently from other play guidance identifiers, so as to prompt the user for a previous location identifier and a next location identifier of the location identifier corresponding to the video clip currently played.
The step 107 may be performed before the step 102, after the step 102, or simultaneously with the step 102. Fig. 7 is merely an example, where step 107 may precede step 102 described above, and embodiments of the present application are not specifically limited.
Therefore, when the first video clip is played, the electronic device can update the display parameters of the first playing guide identifier corresponding to the first place identifier corresponding to the first video clip, so that the first playing guide identifier corresponding to the first place identifier is displayed in other playing guide identifiers in a distinguishing way, and a user can quickly know the currently played video clip and the video clips before and after the currently played video clip. Therefore, the user can quickly learn the change condition of the places in the video, and the efficiency of positioning the video clips is improved.
In some embodiments of the present application, after the step 107, the video playing method provided in the embodiments of the present application may further include the following steps 108 and 109.
Step 108, the electronic device receives a fourth input of the user to the first sub-identifier in the first play guidance identifier.
Step 109, the electronic device displays a prompt message in response to the fourth input.
Wherein, the prompt information has association relation with the video content of the second video clip;
in some embodiments of the present application, the second video clip is a video clip indicated by the first sub-identifier, and the second video clip is a video clip corresponding to the second location identifier, where the location indicated by the second location identifier is different from the location indicated by the first location identifier.
In some embodiments of the present application, in a case where the first location identifier corresponds to a plurality of video clips, the first play guide identifier corresponding to the first location identifier may include a plurality of sub-identifiers indicating play orders of the plurality of video clips and other video clips. The user can trigger the electronic device to display the prompt information through the input of one of the plurality of sub-identifications.
In some embodiments of the present application, the prompting information may include a profile of the video clip corresponding to the first location identifier, so as to prompt the user for video content corresponding to each video clip, so that the user may select the video clip to be watched according to the profiles of the plurality of video clips.
In some embodiments of the present application, the first sub-identifier may be any play guide identifier corresponding to the first location identifier in the first play guide identifiers.
In some embodiments of the present application, the fourth input may be used to trigger the electronic device to display a prompt.
For example, the electronic device may receive a click input of the first sub-identifier from the user and display a prompt.
For example, the electronic device may receive a long press input of the first sub-identifier by the user, and display a prompt message.
Illustratively, taking the first location identifier as the location identifier 1, the first play guide identifier includes the play guide identifier 1 and the play guide identifier 2 as examples. Assuming that the location identifier 1 corresponds to the video segment 1, the location identifier 2 corresponds to the video segment 2, the location identifier 3 corresponds to the video segment 3, and the location identifier 1 corresponds to the play guide identifier 1 and the play guide identifier 2, the play guide identifier 1 indicates that the next video frequency band of the video segment 1 corresponding to the location identifier 1 is the video segment 2 corresponding to the location identifier 2, and the play guide identifier 2 indicates that the next video frequency band of the video segment 1 corresponding to the location identifier 1 is the video segment 3 corresponding to the location identifier 3. Then, the electronic device may receive a fourth input from the user, and display a prompt message having an association with the video content of the second video clip.
Therefore, the electronic equipment can display the prompt information with the association relation with the video content of the video clip, so that a user can quickly know the scenario content of the video clip, the video clip to be watched can be quickly selected, and the positioning efficiency of the video clip is improved.
In some embodiments of the present application, as shown in fig. 9 in conjunction with fig. 7, after the step 107, the video playing method provided in the embodiments of the present application may further include the following steps 110 and 111.
Step 110, the electronic device receives a fifth input of the user to the second sub-identifier in the first play guiding identifier.
And step 111, the electronic device responds to the fifth input, and continues to play the third video clip indicated by the second sub-identifier after the first video clip is played.
The third video clip is a video clip corresponding to a third location identifier, and the location indicated by the third location identifier is different from the location indicated by the first location identifier.
In some embodiments of the present application, the second sub-identifier may be the same as or different from the first sub-identifier.
In some embodiments of the present application, the fifth input is used to select a next video clip to be played after the first video clip is played. In other words, the fifth input may be a selection input of a next played video clip.
For example, the electronic device may receive a click input of the second sub-identifier from the user, and after the first video clip is played, continue playing the third video clip indicated by the second sub-identifier.
For example, the electronic device may receive a long press input from the user to the second sub-identifier, and after the first video clip is played, continue playing the third video clip indicated by the second sub-identifier.
Thus, the user can flexibly select the video clip to be played next through the input of the second sub-identifier in the first playing guide identifier. Thus, the flexibility of video playing is improved.
In some embodiments of the present application, after the step 107, the video playing method provided in the embodiments of the present application may further include the following steps 112 and 113.
Step 112, the electronic device receives a sixth input of the first playing guidance identifier from the user.
Step 113, the electronic device sets a playing sequence of at least two fourth video clips after playing the first video clip in response to the sixth input.
The at least two fourth video clips are video clips indicated by the first playing guide identifier, and each fourth video clip corresponds to a different location identifier.
In some embodiments of the present application, the above "play order of at least two fourth video clips after playing the first video clip" may be understood as: the electronic device sets whether the at least two fourth video clips are played after the first video clip. That is, through the sixth input of the user, the electronic device can fixedly play the video clip indicated by the first play guide identifier, without playing the video clip indicated by the other play guide identifier.
Illustratively, the first playhead guide identification indicates video clip 1 and video clip 2. The electronic device may receive a double-click input of the user on the first playing guidance identifier, and set the playing order of the video clip 1 and the video clip 2 to play the video clip 1 first and then play the video clip 2 after the first video clip.
Therefore, the electronic equipment can determine the played video clips according to the input of the playing guide identifier by the user, so that the quick selection of the video clips is realized, and the positioning efficiency of the video clips is improved.
In some embodiments of the present application, the video playing method provided in the embodiments of the present application may further include the following steps 114 and 115.
Step 114, the electronic device receives a seventh input of the user when displaying at least one location identifier corresponding to the first video.
In response to the seventh input, the electronic device displays at least one object identifier or at least one scenario identifier in a corresponding area of a fourth location identifier of the at least one location identifier 115.
Wherein, each object identifier indicates an object in the first video, the video segment corresponding to the fourth location identifier includes at least one object indicated by the object identifier, and each scenario identifier has an association relationship with the video content of the video segment corresponding to the fourth location identifier.
Illustratively, take the first video as an example of a television show. The object identifier may be a character identifier of a lead actor.
Illustratively, the first video is exemplified as an animation film. The object identifier may be an identifier of a principal angle.
It will be appreciated that the electronic device may enable a user to quickly learn about objects appearing in the video clip corresponding to the fourth location identifier by displaying at least one object identifier.
In some embodiments of the present application, the electronic device may analyze, by using an AI method, subtitles, pictures and audio of a video clip corresponding to the first video, to obtain an object included in the first video, and associate the object with the first location identifier according to a time when the object appears.
In some embodiments of the present application, the scenario identifier is scenario information for indicating a video clip corresponding to the fourth location identifier.
In some embodiments of the present application, the scenario information may be information describing video content of the video clip corresponding to the fourth location identifier.
For example, assuming that the scenario information indicated by the scenario identification is a horror scenario, the video clip corresponding to the fourth location identification is a horror video clip.
For example, assuming that the scenario information indicated by the scenario identification is a confusing scenario, the fourth location identification corresponds to a confusing video clip.
It can be appreciated that the electronic device can display at least one scenario identifier, so that the user can quickly know the rough content of the video clip through the at least one scenario identifier, and the user can determine whether to watch the video clip in advance.
In some embodiments of the present application, the corresponding area of the fourth location identifier may be an area having an association relationship with a display area of the fourth location identifier.
For example, the corresponding region of the fourth location identifier may be a display region of the fourth location identifier.
The corresponding area of the fourth location identifier may also be, for example, an area surrounding the display area of the first location identifier.
In some embodiments of the present application, the fourth location identifier may be the same as or different from the first location identifier.
In some embodiments of the present application, the electronic device may receive, when playing the video clip corresponding to the fourth location identifier, an input that the user annotates the video clip, and display the corresponding scenario identifier in association with the fourth location identifier.
Illustratively, as shown in FIG. 10, the electronic device can receive user input to the annotation control 101 to initiate functionality for identifying an associated scenario for a place. Then, the electronic device may receive an input from the user to add a scenario identifier to the location identifier a, and add a corresponding scenario identifier to the location identifier a.
In some embodiments of the present application, the electronic device may further analyze, through AI, video content or a barrage of a video clip corresponding to the fourth location identifier, and automatically add scenario identifiers and object identifiers to the video clip. Thus, the user can quickly know the approximate scenario of the video clip through scenario identification and object identification without watching the video clip.
Therefore, the user can quickly know the approximate scenario of the video segment through the object identifier or the scenario identifier because the electronic equipment can display at least one of the object identifier or the scenario identifier associated with the fourth place identifier, so that the video segment can be quickly screened, and the positioning efficiency of the video segment is improved.
In some embodiments of the present application, the electronic device further receives user input of a fourth location identifier and a first object identifier, and plays the target video clip.
Wherein, the first object identifier may be one or more of at least one object identifier; the target video clip may be: the fourth location identifier corresponds to at least one video clip, and the video clip contains the object indicated by the first object identifier.
In some embodiments of the present application, the electronic device may receive an input from a user on the fourth location identifier and the first object identifier, and play the video segment of the object corresponding to the first object identifier in the at least one video segment corresponding to the fourth location identifier. Thus, the video clips can be screened more accurately.
Illustratively, as shown in FIG. 10, the electronic device can receive user input to the object selection control 102 after the user selects the place identifier B, displaying the object identifier 103, the object identifier 104, and the object identifier 105. Then, the electronic device may receive a user selection input of the object identifier 103, and play the video clip including the object identifier 103 in the at least one video clip corresponding to the location identifier B.
Thus, the user can select the object in the video clip on the basis of the selection of the location identifier corresponding to the video clip. Therefore, the video clips can be screened more accurately through double selection of the location identifier and the object identifier corresponding to the video clips, and the positioning efficiency of the video clips is improved.
The above method embodiments, or various possible implementation manners in the method embodiments, may be executed separately, or may be executed in any two or more combinations, which may be specifically determined according to actual use requirements, and this embodiment of the application is not limited.
According to the video playing method provided by the embodiment of the application, the execution main body can be a video playing device. In the embodiment of the present application, a video playing device executes a video playing method as an example, and the video playing device provided in the embodiment of the present application is described.
Fig. 11 shows a schematic diagram of a possible structure of a video playing device according to an embodiment of the present application. As shown in fig. 11, the video playback device 110 may include: a display module 111, a receiving module 112 and a processing module 113.
The receiving module 112 is configured to receive a first input of a user to a first location identifier in the at least one location identifier when the display module 111 displays the at least one location identifier corresponding to the first video; the processing module 113 is configured to play a first video clip corresponding to the first location identifier in response to the first input received by the receiving module 112; wherein the first video comprises at least two video clips, one location identifier corresponds to the at least one video clip, and one location identifier indicates one location in the first video.
In a possible implementation manner, the receiving module 112 is further configured to receive a second input of the user before receiving the first input of the first location identifier of the at least one location identifier;
the display module 111 is further configured to display a video map corresponding to the first video in response to the second input received by the receiving module 112, where the video map includes at least one location identifier.
In a possible implementation manner, the receiving module 112 is further configured to receive a third input of the user after the display module 111 displays the video map corresponding to the first video;
the processing module 113 is further configured to update a display parameter of the video map in response to the third input received by the receiving module 112;
wherein the display parameters include at least one of: display position, display size, display shape, display color, display position of at least one location indicator, name of at least one location indicator.
In one possible implementation manner, the video map includes at least two play guide identifiers, where the play guide identifiers are used to indicate at least two video clips and a play sequence of the at least two video clips;
the processing module 113 is further configured to update, after the receiving module 112 receives a first input of a first location identifier from the at least one location identifier, a display parameter of a first play guide identifier corresponding to the first location identifier in response to the first input.
In a possible implementation manner, the receiving module 112 is further configured to receive a fourth input of the first sub-identifier in the first play guiding identifier from the user after the processing module 113 updates the display parameter of the first play guiding identifier corresponding to the first location identifier;
the display module 111 is further configured to display a prompt message in response to the fourth input received by the receiving module 112, where the prompt message has an association relationship with the video content of the second video segment;
the second video clip is a video clip indicated by the first sub-identifier, and the second video clip is a video clip corresponding to the second location identifier, where the location indicated by the second location identifier is different from the location indicated by the first location identifier.
In a possible implementation manner, the receiving module 112 is further configured to receive a fifth input of the second sub-identifier in the first play guiding identifier from the user after the processing module 113 updates the display parameter of the first play guiding identifier corresponding to the first location identifier;
the processing module 113 is further configured to, in response to the fifth input received by the receiving module 112, continue playing the third video segment indicated by the second sub-identifier after the first video segment is played;
The third video clip is a video clip corresponding to a third location identifier, and the location indicated by the third location identifier is different from the location indicated by the first location identifier.
In a possible implementation manner, the receiving module 112 is further configured to receive a sixth input of the first play guidance identifier from the user after the processing module 113 updates the display parameter of the first play guidance identifier corresponding to the first location identifier;
the processing module 113 is further configured to set a playing order of at least two fourth video clips after the first video clip is played in response to the sixth input received by the receiving module 112;
the at least two fourth video clips are video clips indicated by the first playing guide identifier, and each fourth video clip corresponds to a different place identifier.
In a possible implementation manner, the receiving module 112 is further configured to receive a seventh input of the user if at least one location identifier corresponding to the first video is displayed;
the display module 111 is further configured to display at least one object identifier or at least one scenario identifier in a corresponding area of a fourth location identifier in the at least one location identifier in response to the seventh input received by the receiving module 112;
Each object identifier indicates an object in the first video, the video segment corresponding to the fourth location identifier includes at least one object indicated by the object identifier, and each scenario identifier has an association relationship with video content of the video segment corresponding to the fourth location identifier.
The embodiment of the application provides a video playing device, which can display a place identifier corresponding to a place in a video, and a user selects the place identifier of interest to play a video clip corresponding to the selected place identifier.
The video playing device in the embodiment of the application may be an electronic device, or may be a component in an electronic device, for example, an integrated circuit or a chip. The electronic device may be a terminal, or may be other devices than a terminal. By way of example, the electronic device may be a mobile phone, tablet computer, notebook computer, palm computer, vehicle-mounted electronic device, mobile internet appliance (Mobile Internet Device, MID), augmented reality (augmented reality, AR)/Virtual Reality (VR) device, robot, wearable device, ultra-mobile personal computer, UMPC, netbook or personal digital assistant (personal digital assistant, PDA), etc., but may also be a server, network attached storage (Network Attached Storage, NAS), personal computer (personal computer, PC), television (TV), teller machine or self-service machine, etc., and the embodiments of the present application are not limited in particular.
The video playing device in the embodiment of the present application may be a device having an operating system. The operating system may be an Android operating system, an ios operating system, or other possible operating systems, which are not specifically limited in the embodiments of the present application.
The video playing device provided in the embodiment of the present application can implement each process implemented by the embodiment of the video playing method, so as to achieve the same technical effect, and in order to avoid repetition, a detailed description is omitted here.
Optionally, as shown in fig. 12, the embodiment of the present application further provides an electronic device 1200, including a processor 1201 and a memory 1202, where the memory 1202 stores a program or an instruction that can be executed on the processor 1201, and the program or the instruction when executed by the processor 1201 implements each step of the embodiment of the video playing method, and can achieve the same technical effect, so that repetition is avoided, and no further description is given here.
The electronic device in the embodiment of the application includes the mobile electronic device and the non-mobile electronic device described above.
Fig. 13 is a schematic hardware structure of an electronic device implementing an embodiment of the present application.
The electronic device 1300 includes, but is not limited to: radio frequency unit 1301, network module 1302, audio output unit 1303, input unit 1304, sensor 1305, display unit 1306, user input unit 1307, interface unit 1308, memory 1309, and processor 1310.
Those skilled in the art will appreciate that the electronic device 1300 may also include a power source (e.g., a battery) for powering the various components, which may be logically connected to the processor 1310 by a power management system, such as to perform functions such as managing charging, discharging, and power consumption by the power management system. The electronic device structure shown in fig. 13 does not constitute a limitation of the electronic device, and the electronic device may include more or less components than shown, or may combine certain components, or may be arranged in different components, which are not described in detail herein.
Wherein, the user input unit 1307 is configured to receive, when the display unit 1306 displays at least one location identifier corresponding to the first video, a first input of a user to a first location identifier in the at least one location identifier; a processor 1310, configured to play a first video clip corresponding to the first location identifier in response to the first input received by the user input unit 1307; wherein the first video comprises at least two video clips, one location identifier corresponds to the at least one video clip, and one location identifier indicates one location in the first video.
In a possible implementation manner, the user input unit 1307 is further configured to receive a second input by the user before receiving the first input by the user for the first location identifier in the at least one location identifier;
The display unit 1306 is further configured to display a video map corresponding to the first video, where the video map includes at least one location identifier, in response to the second input received by the user input unit 1307.
In a possible implementation manner, the user input unit 1307 is further configured to receive a third input from the user after the display unit 1306 displays the video map corresponding to the first video;
the processor 1310 is further configured to update a display parameter of the video map in response to the third input received by the user input unit 1307;
wherein the display parameters include at least one of: display position, display size, display shape, display color, display position of at least one location indicator, name of at least one location indicator.
In one possible implementation manner, the video map includes at least two play guide identifiers, where the play guide identifiers are used to indicate at least two video clips and a play sequence of the at least two video clips;
the above-mentioned processor 1310 is further configured to update, after the user input unit 1307 receives a first input of a first location identifier from the at least one location identifier, a display parameter of a first play guide identifier corresponding to the first location identifier in response to the first input.
In a possible implementation manner, the user input unit 1307 is further configured to receive a fourth input of the first sub-identifier in the first play guidance identifier by the user after the processor 1310 updates the display parameter of the first play guidance identifier corresponding to the first location identifier;
the display unit 1306 is further configured to display, in response to the fourth input received by the user input unit 1307, a prompt message, where the prompt message has an association relationship with the video content of the second video clip;
the second video clip is a video clip indicated by the first sub-identifier, and the second video clip is a video clip corresponding to the second location identifier, where the location indicated by the second location identifier is different from the location indicated by the first location identifier.
In a possible implementation manner, the user input unit 1307 is further configured to receive a fifth input of the second sub-identifier in the first play guidance identifier by the user after the processor 1310 updates the display parameter of the first play guidance identifier corresponding to the first location identifier;
the processor 1310 is further configured to, in response to the fifth input received by the user input unit 1307, continue playing the third video clip indicated by the second sub-identifier after the first video clip is played;
The third video clip is a video clip corresponding to a third location identifier, and the location indicated by the third location identifier is different from the location indicated by the first location identifier.
In a possible implementation manner, the user input unit 1307 is further configured to receive a sixth input of the first play guidance identifier by the user after the processor 1310 updates the display parameter of the first play guidance identifier corresponding to the first location identifier;
the processor 1310 is further configured to set a playing order of at least two fourth video clips after the first video clip is played in response to the sixth input received by the user input unit 1307;
the at least two fourth video clips are video clips indicated by the first playing guide identifier, and each fourth video clip corresponds to a different place identifier.
In a possible implementation manner, the user input unit 1307 is further configured to receive a seventh input of the user if at least one location identifier corresponding to the first video is displayed;
the display unit 1306 is further configured to display at least one object identifier or at least one scenario identifier in a corresponding area of a fourth location identifier in the at least one location identifier in response to the seventh input received by the user input unit 1307;
Each object identifier indicates an object in the first video, the video segment corresponding to the fourth location identifier includes at least one object indicated by the object identifier, and each scenario identifier has an association relationship with video content of the video segment corresponding to the fourth location identifier.
The embodiment of the application provides electronic equipment, which can display the place identifier corresponding to the place in the video, and the user selects the place identifier of interest to play the video clip corresponding to the selected place identifier.
It should be appreciated that in embodiments of the present application, the input unit 1304 may include a graphics processor (Graphics Processing Unit, GPU) 13041 and a microphone 13042, the graphics processor 13041 processing image data of still pictures or video obtained by an image capturing device (e.g., a camera) in a video capturing mode or an image capturing mode. The display unit 1306 may include a display panel 13061, and the display panel 13061 may be configured in the form of a liquid crystal display, an organic light emitting diode, or the like. The user input unit 1307 includes at least one of a touch panel 13071 and other input devices 13072. The touch panel 13071 is also referred to as a touch screen. The touch panel 13071 can include two parts, a touch detection device and a touch controller. Other input devices 13072 may include, but are not limited to, a physical keyboard, function keys (e.g., volume control keys, switch keys, etc.), a trackball, a mouse, a joystick, and so forth, which are not described in detail herein.
Memory 1309 may be used to store software programs as well as various data. The memory 1309 may mainly include a first memory area storing programs or instructions and a second memory area storing data, wherein the first memory area may store an operating system, application programs or instructions (such as a sound playing function, an image playing function, etc.) required for at least one function, and the like. Further, the memory 1309 may include volatile memory or nonvolatile memory, or the memory 1309 may include both volatile and nonvolatile memory. The nonvolatile Memory may be a Read-Only Memory (ROM), a Programmable ROM (PROM), an Erasable PROM (EPROM), an Electrically Erasable EPROM (EEPROM), or a flash Memory. The volatile memory may be random access memory (Random Access Memory, RAM), static RAM (SRAM), dynamic RAM (DRAM), synchronous DRAM (SDRAM), double Data Rate SDRAM (ddr SDRAM), enhanced SDRAM (Enhanced SDRAM), synchronous DRAM (SLDRAM), and Direct RAM (DRRAM). Memory 1309 in embodiments of the present application includes, but is not limited to, these and any other suitable types of memory.
The processor 1310 may include one or more processing units; optionally, processor 1310 integrates an application processor that primarily handles operations related to the operating system, user interface, and applications, and a modem processor that primarily handles wireless communication signals, such as a baseband processor. It will be appreciated that the modem processor described above may not be integrated into the processor 1310.
The embodiment of the present application further provides a readable storage medium, where a program or an instruction is stored on the readable storage medium, and when the program or the instruction is executed by a processor, the program or the instruction realizes each process of the embodiment of the video playing method, and the same technical effects can be achieved, so that repetition is avoided, and no description is repeated here.
Wherein the processor is a processor in the electronic device described in the above embodiment. The readable storage medium includes computer readable storage medium such as computer readable memory ROM, random access memory RAM, magnetic or optical disk, etc.
The embodiment of the application further provides a chip, the chip includes a processor and a communication interface, the communication interface is coupled with the processor, the processor is used for running a program or instructions, implementing each process of the video playing method embodiment, and achieving the same technical effect, so as to avoid repetition, and no redundant description is provided herein.
It should be understood that the chips referred to in the embodiments of the present application may also be referred to as system-on-chip chips, chip systems, or system-on-chip chips, etc.
The embodiments of the present application provide a computer program product stored in a storage medium, where the program product is executed by at least one processor to implement the respective processes of the embodiments of the video playing method described above, and achieve the same technical effects, and are not repeated herein.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising one … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element. Furthermore, it should be noted that the scope of the methods and apparatus in the embodiments of the present application is not limited to performing the functions in the order shown or discussed, but may also include performing the functions in a substantially simultaneous manner or in an opposite order depending on the functions involved, e.g., the described methods may be performed in an order different from that described, and various steps may also be added, omitted, or combined. Additionally, features described with reference to certain examples may be combined in other examples.
From the above description of the embodiments, it will be clear to those skilled in the art that the above-described embodiment method may be implemented by means of software plus a necessary general hardware platform, but of course may also be implemented by means of hardware, but in many cases the former is a preferred embodiment. Based on such understanding, the technical solutions of the present application may be embodied essentially or in a part contributing to the prior art in the form of a computer software product stored in a storage medium (such as ROM/RAM, magnetic disk, optical disk), comprising several instructions for causing a terminal (which may be a mobile phone, a computer, a server, or a network device, etc.) to perform the methods described in the embodiments of the present application.
The embodiments of the present application have been described above with reference to the accompanying drawings, but the present application is not limited to the above-described embodiments, which are merely illustrative and not restrictive, and many forms may be made by those of ordinary skill in the art without departing from the spirit of the present application and the scope of the claims, which are also within the protection of the present application.

Claims (18)

1. A video playing method, the method comprising:
receiving a first input of a user to a first place identifier in at least one place identifier under the condition that the at least one place identifier corresponding to the first video is displayed;
responsive to the first input, playing a first video clip corresponding to the first location identifier;
wherein the first video comprises at least two video clips, one of the place identifiers corresponds to at least one of the video clips, and one of the place identifiers indicates one place in the first video.
2. The method of claim 1, wherein prior to receiving the first user input for a first one of the at least one location identifier, the method further comprises:
receiving a second input from the user;
and in response to the second input, displaying a video map corresponding to the first video, the video map including the at least one location identifier.
3. The method of claim 2, wherein after the displaying the video map corresponding to the first video, the method further comprises:
receiving a third input from the user;
In response to the third input, updating display parameters of the video map;
wherein the display parameters include at least one of: a display position, a display size, a display shape, a display color, a display position of at least one of the place marks, a name of at least one of the place marks.
4. The method of claim 2, wherein the video map comprises at least two playguide identifiers indicating at least two video clips and a playorder of the at least two video clips;
after the receiving a first input by the user to a first location identifier of the at least one location identifier, the method further comprises:
and in response to the first input, updating display parameters of a first play guide identifier corresponding to the first place identifier.
5. The method of claim 4, wherein after updating the display parameters of the first playguide identifier corresponding to the first location identifier, the method further comprises:
receiving a fourth input of a user to a first sub-identifier in the first playing guide identifier;
responding to the fourth input, and displaying prompt information, wherein the prompt information has an association relationship with the video content of the second video clip;
The second video segment is a video segment indicated by the first sub-identifier, and the second video segment is a video segment corresponding to a second location identifier, where the location indicated by the second location identifier is different from the location indicated by the first location identifier.
6. The method of claim 4, wherein after updating the display parameters of the first playguide identifier corresponding to the first location identifier, the method further comprises:
receiving a fifth input of a user to a second sub-identifier in the first playing guide identifier;
responding to the fifth input, and continuing to play a third video clip indicated by the second sub-identifier after the first video clip is played;
the third video segment is a video segment corresponding to a third place identifier, and the place indicated by the third place identifier is different from the place indicated by the first place identifier.
7. The method of claim 4, wherein after updating the display parameters of the first playguide identifier corresponding to the first location identifier, the method further comprises:
receiving a sixth input of a user to the first playing guide identifier;
Setting the playing sequence of at least two fourth video clips after the first video clip is played in response to the sixth input;
the at least two fourth video clips are video clips indicated by the first playing guide identifier, and each fourth video clip corresponds to a different place identifier.
8. The method of claim 1, wherein in the case of displaying at least one location identifier corresponding to the first video, the method further comprises:
receiving a seventh input of the user;
in response to the seventh input, displaying at least one object identifier or at least one scenario identifier in a corresponding area of a fourth location identifier of the at least one location identifier;
each object identifier indicates an object in the first video, the object indicated by the at least one object identifier is included in the video segment corresponding to the fourth location identifier, and each scenario identifier has an association relationship with video content of the video segment corresponding to the fourth location identifier.
9. A video playback device, the device comprising: the device comprises a display module, a receiving module and a processing module;
The receiving module is used for receiving a first input of a user to a first place identifier in the at least one place identifier under the condition that the display module displays the at least one place identifier corresponding to the first video;
the processing module is used for responding to the first input received by the receiving module and playing a first video clip corresponding to the first place identifier;
wherein the first video comprises at least two video clips, one of the place identifiers corresponds to at least one of the video clips, and one of the place identifiers indicates one place in the first video.
10. The apparatus of claim 9, wherein the receiving module is further configured to receive a second input from the user prior to receiving the first input from the user to a first one of the at least one location identifier;
the display module is further configured to display a video map corresponding to the first video in response to the second input received by the receiving module, where the video map includes the at least one location identifier.
11. The apparatus of claim 10, wherein the receiving module is further configured to receive a third input from a user after the display module displays a video map corresponding to the first video;
The processing module is further used for responding to the third input received by the receiving module and updating the display parameters of the video map;
wherein the display parameters include at least one of: a display position, a display size, a display shape, a display color, a display position of at least one of the place marks, a name of at least one of the place marks.
12. The apparatus of claim 10, wherein the video map comprises at least two playguide identifiers indicating at least two video clips and a playorder of the at least two video clips;
the processing module is further configured to update, after the receiving module receives a first input of a first location identifier from the at least one location identifier, a display parameter of a first play guidance identifier corresponding to the first location identifier in response to the first input.
13. The apparatus of claim 12, wherein the receiving module is further configured to receive a fourth input from a user of a first sub-identifier in the first playguide identifier after the processing module updates display parameters of the first playguide identifier corresponding to the first location identifier;
The display module is further configured to display a prompt message in response to the fourth input received by the receiving module, where the prompt message has an association relationship with video content of the second video segment;
the second video segment is a video segment indicated by the first sub-identifier, and the second video segment is a video segment corresponding to a second location identifier, where the location indicated by the second location identifier is different from the location indicated by the first location identifier.
14. The apparatus of claim 12, wherein the receiving module is further configured to receive a fifth input from a user of a second sub-identifier in the first playguide identifier after the processing module updates the display parameters of the first playguide identifier corresponding to the first location identifier;
the processing module is further configured to, in response to the fifth input received by the receiving module, continue playing the third video segment indicated by the second sub-identifier after the first video segment is played;
the third video segment is a video segment corresponding to a third place identifier, and the place indicated by the third place identifier is different from the place indicated by the first place identifier.
15. The apparatus of claim 12, wherein the receiving module is further configured to receive a sixth input of the first playguide identifier from the user after the processing module updates the display parameters of the first playguide identifier corresponding to the first location identifier;
the processing module is further configured to set a playing sequence of at least two fourth video clips after the first video clip is played in response to the sixth input received by the receiving module;
the at least two fourth video clips are video clips indicated by the first playing guide identifier, and each fourth video clip corresponds to a different place identifier.
16. The apparatus of claim 9, wherein the receiving module is further configured to receive a seventh input from the user if the at least one location identifier corresponding to the first video is displayed;
the display module is further configured to display at least one object identifier or at least one scenario identifier in a corresponding area of a fourth location identifier in the at least one location identifier in response to the seventh input received by the receiving module;
Each object identifier indicates an object in the first video, the object indicated by the at least one object identifier is included in the video segment corresponding to the fourth location identifier, and each scenario identifier has an association relationship with video content of the video segment corresponding to the fourth location identifier.
17. An electronic device comprising a processor and a memory storing a program or instructions executable on the processor, which when executed by the processor, implement the steps of the video playback method of any one of claims 1 to 8.
18. A readable storage medium, wherein a program or instructions is stored on the readable storage medium, which when executed by a processor, implements the steps of the video playback method of any one of claims 1 to 8.
CN202311385736.0A 2023-10-24 2023-10-24 Video playing method and device, electronic equipment and readable storage medium Pending CN117294900A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311385736.0A CN117294900A (en) 2023-10-24 2023-10-24 Video playing method and device, electronic equipment and readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311385736.0A CN117294900A (en) 2023-10-24 2023-10-24 Video playing method and device, electronic equipment and readable storage medium

Publications (1)

Publication Number Publication Date
CN117294900A true CN117294900A (en) 2023-12-26

Family

ID=89253415

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311385736.0A Pending CN117294900A (en) 2023-10-24 2023-10-24 Video playing method and device, electronic equipment and readable storage medium

Country Status (1)

Country Link
CN (1) CN117294900A (en)

Similar Documents

Publication Publication Date Title
US11417367B2 (en) Systems and methods for reviewing video content
KR20160024002A (en) Method for providing visual sound image and electronic device implementing the same
CN112672061B (en) Video shooting method and device, electronic equipment and medium
CN113905175A (en) Video generation method and device, electronic equipment and readable storage medium
KR20130124866A (en) Mobile terminal and method for controlling thereof
CN113918522A (en) File generation method and device and electronic equipment
CN113596555B (en) Video playing method and device and electronic equipment
CN112887794B (en) Video editing method and device
CN111679772B (en) Screen recording method and system, multi-screen device and readable storage medium
EP4343579A1 (en) Information replay method and apparatus, electronic device, computer storage medium, and product
WO2022179415A1 (en) Audiovisual work display method and apparatus, and device and medium
CN113852757B (en) Video processing method, device, equipment and storage medium
CN113096686B (en) Audio processing method and device, electronic equipment and storage medium
CN115437736A (en) Method and device for recording notes
CN112367487B (en) Video recording method and electronic equipment
CN114679546A (en) Display method and device, electronic equipment and readable storage medium
CN117294900A (en) Video playing method and device, electronic equipment and readable storage medium
CN113923392A (en) Video recording method, video recording device and electronic equipment
CN114302009A (en) Video processing method, video processing device, electronic equipment and medium
US10637905B2 (en) Method for processing data and electronic apparatus
CN114390205B (en) Shooting method and device and electronic equipment
CN114500852B (en) Shooting method, shooting device, electronic equipment and readable storage medium
CN114500852A (en) Photographing method, photographing apparatus, electronic device, and readable storage medium
CN117149038A (en) Image display method and image display device
CN114860122A (en) Application program control method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination