CN117097937A - Map generation method, device, equipment and medium - Google Patents

Map generation method, device, equipment and medium Download PDF

Info

Publication number
CN117097937A
CN117097937A CN202210520977.0A CN202210520977A CN117097937A CN 117097937 A CN117097937 A CN 117097937A CN 202210520977 A CN202210520977 A CN 202210520977A CN 117097937 A CN117097937 A CN 117097937A
Authority
CN
China
Prior art keywords
map
video file
file
position information
video
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210520977.0A
Other languages
Chinese (zh)
Inventor
方建伟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Pateo Network Technology Service Co Ltd
Original Assignee
Shanghai Pateo Network Technology Service Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Pateo Network Technology Service Co Ltd filed Critical Shanghai Pateo Network Technology Service Co Ltd
Priority to CN202210520977.0A priority Critical patent/CN117097937A/en
Publication of CN117097937A publication Critical patent/CN117097937A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • H04N21/4312Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
    • H04N21/4316Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations for displaying supplemental content in a region of the screen, e.g. an advertisement in a separate window
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/787Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using geographical or spatial information, e.g. location

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Signal Processing (AREA)
  • Library & Information Science (AREA)
  • Business, Economics & Management (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Marketing (AREA)
  • Instructional Devices (AREA)

Abstract

The invention discloses a map generation method, a map generation device, map generation equipment and map generation media; the map generation method comprises the following steps: obtaining geographic position information in a video file; generating an auxiliary map according to the geographic position information; and synchronously displaying the auxiliary map when playing the video file. According to the map generation method, the map generation device, the map generation equipment and the map generation medium, the auxiliary map is displayed based on the geographic position information in the video file, and the auxiliary map is synchronously displayed when the video file is played, so that a user can know the geographic position in the video in time in the watching process, understanding and impression of the watched video are facilitated, and video watching experience is improved.

Description

Map generation method, device, equipment and medium
Technical Field
The present invention relates to the field of multimedia technologies, and in particular, to a map generating method and apparatus, a device, and a medium.
Background
When we watch a video work, many geographical locations are usually involved in the film, and as the video picture is switched and the scenario progresses, the places or scenes jump continuously, and the audience cannot know the actual distribution of the relevant geographical locations while watching the video. For example, may miss some impressive places, resulting in distraction or regretation that affects the user's experience of viewing the video. For many places mentioned in the video, the video is recorded in time due to no time in the watching process, and the video is inconvenient to turn over after being watched.
Disclosure of Invention
The invention aims to overcome the defect that a user cannot know the geographic position in a video in time when watching the video in the prior art, and provides a map generation method, a map generation device, map generation equipment and map generation media.
The invention solves the technical problems by the following technical scheme:
the invention provides a map generation method, which comprises the following steps:
obtaining geographic position information in the video file;
generating an auxiliary map according to the geographic position information;
and synchronously displaying the auxiliary map when the video file is played.
The invention also provides a map generation device, which comprises a display unit, one or more processing units and a storage unit, wherein the one or more processing units are respectively in communication connection with the storage unit and the display unit;
the storage unit is configured to store instructions that, when executed by the one or more processing units, cause the one or more processing units to perform steps comprising:
obtaining geographic position information in the video file;
generating an auxiliary map according to the geographic position information;
and synchronously displaying the auxiliary map when the display unit plays the video file.
The invention also provides an electronic device, which comprises a memory, a processor and a computer program stored on the memory and capable of running on the processor, wherein the map generation method is realized when the processor executes the computer program.
The present invention also provides a computer-readable storage medium having stored thereon a computer program which, when executed by a processor, implements the map generation method described above.
The invention has the positive progress effects that: according to the map generation method, the map generation device, the map generation equipment and the map generation medium, the auxiliary map is displayed based on the geographic position information in the video file, and the auxiliary map is synchronously displayed when the video file is played, so that a user can know the geographic position in the video in time in the watching process, understanding and impression of the watched video are facilitated, and video watching experience is improved.
Drawings
Fig. 1 is a flowchart of a map generation method in embodiment 1 of the present invention.
Fig. 2 is an application example diagram of the map generation method in embodiment 1 of the present invention.
Fig. 3 is a schematic block diagram of a map generating apparatus in embodiment 2 of the present invention.
Fig. 4 is a block diagram showing the structure of an electronic device in embodiment 3 of the invention
Detailed Description
The invention is further illustrated by means of the following examples, which are not intended to limit the scope of the invention.
Example 1
Referring to fig. 1, the embodiment specifically provides a map generating method, which includes:
s1, obtaining geographic position information in a video file;
s2, generating an auxiliary map according to the geographic position information;
s3, synchronously displaying the auxiliary map when the video file is played.
The formats of the video file in step S1 include, but are not limited to, AVI (Audio Video Interleaved, audio video interleave format), WMV (Windows Media Video, microsoft video codec format), MPEG (Moving Picture Experts Group ), MP4 (Moving Picture Experts Group, moving picture experts group 4 generation format), FLV (FlashVideo, streaming media format), ASF (Advanced Streaming Format ), 3GPP (3 rd Generation Partnership Project, third generation partnership project 3G streaming media video coding container format), VOB (Video Object file format, multimedia container format), and the like.
The geographical location information in the video file in step S1 is related information for describing the characteristics of the geographical objects. The geographic location information in the video file may be expressed in the form of media of different dimensions, including but not limited to, visual and audio content in the video file. In step S1, the geographical location information in the video file may be obtained for different media, for example, when the user views the video file in an APP (application program) of a certain terminal, the content transferred by the video file may be obtained from different dimensions as long as the APP is authorized to have corresponding rights such as system sound, screen capturing, microphone pickup, etc., and further the geographical location information therein may be obtained by means of OCR (Optical Character Recognition, optical text recognition) and the like.
The auxiliary map in step S2, i.e. the map for assisting the user to watch the video, is usually presented in the form of an electronic map on the same carrier as the video file, and may of course be presented in a separate device, a separate screen or in different software. The geographical position information is displayed on the auxiliary map, specifically, the geographical position information can be realized by adding map elements corresponding to the geographical position information in a related map layer, for example, a place name analysis result of 'Sanqingshan' is obtained from a certain video, a corresponding map interest point file is generated, and the map interest point file is loaded in the auxiliary map. Of course, depending on the type of geographic location information, map elements may be presented in a dot, line, plane or other form such as thermodynamic diagram, for example, highlighting a lake or flashing a railway line, etc., and the map types include, but are not limited to, two-dimensional plan, three-dimensional map, indoor map, real-scene map, etc., and may be seamlessly switched depending on the type of geographic location information. Step S3, by synchronously displaying the auxiliary map when the video file is played, a user can know the geographic position mentioned in the video timely, and normal watching progress is not influenced. The linkage presentation of the video file and the auxiliary map can be respectively carried out in the display areas of one or more devices, for example, video is played on a first folding screen of a certain folding screen device, and the auxiliary map comprising geographic position information is presented on a second folding screen; or respectively playing video files on the screen of a certain mobile phone terminal, and displaying the auxiliary map on the screen of the mobile phone terminal through a transparent and draggable floating window.
As an alternative embodiment, step S1 includes:
performing target detection processing or image segmentation processing on the video file to obtain text information corresponding to the video file;
geographic location information is extracted from the textual information.
For geographical location information appearing in some video pictures, such as inscriptions on a city wall or a tablet, the conventional OCR character recognition effect is often poor because the geographical location information is fused in the video pictures and the location arrangement is generally irregular. In this regard, step S1 may also extract text information from the video file based on the object detection algorithm or the image segmentation algorithm. Specifically, the selection may be based on the picture definition and the frame rate of the video file, directly processing the video file in the above manner, or performing the above detection extraction processing on the frame picture of the video file. For example, when the definition of the picture is generally that the pixel value of the video is not high, the latter method may be adopted to perform object detection or image segmentation processing on the frame picture of the video file, extract text information in the form of Chinese characters, letters, numbers, symbols, etc., which include but are not limited to, appearing in the frame picture, and perform semantic analysis based on the text information to extract geographical location information.
The object detection of the frame picture can be understood as comprehensive processing of object identification and object positioning in the picture, namely, not only the classification of the object is required to be identified, but also the specific position of the object in the picture is also required to be obtained, and the object detection can be realized based on a deep learning detection model. Deep learning detection models are generally divided into two types, namely object recognition and object localization by two steps, and end-to-end implementation. The former includes but is not limited to target detection models such as R-CNN, fast-RCNN and the like, and is characterized by low recognition error rate and low missing recognition rate, and the latter includes but is not limited to target detection models such as Yolo, SSD, yoloV and the like; the method has the characteristics of high recognition speed, high accuracy and capability of meeting the real-time requirement.
The image segmentation algorithm for the frame picture can also extract text information from the frame picture, generally adjusts and scales the frame picture to a size meeting the requirement, and inputs a pre-trained recognition model. Taking a full convolution network as an example, obtaining a corresponding feature map according to a preset extraction mode for an input frame picture, and processing the feature map through a specific convolution branch of a full convolution network model to obtain the text information. It will be appreciated by those skilled in the art that the extraction of the Chinese information in the frame is achieved in the above manner, the deep learning detection model used is trained in advance, and the training data may be a plurality of corresponding images and words grouped by geographic information location.
The acquisition of the geographic position information in the video file can be realized by the mode of the target detection algorithm or the image segmentation algorithm. Specifically, text information on a city wall in a video picture can be obtained by processing a Yolo target detection model or a full convolution network model, and corresponding geographic position information is obtained as "Tong guan" through semantic recognition. According to the method, the geographic position information in the video file is obtained by adopting the target detection algorithm or the image segmentation algorithm, and the possibly irregular text information can be obtained from a relatively complex video picture, so that the geographic position information can be accurately obtained.
As an alternative embodiment, step S1 includes:
performing image matching processing on at least one frame of the video file;
and determining geographic position information according to the image matching processing result.
In this embodiment, the image matching process is directly performed on the frame of the video file, so that the geographical location information contained in the frame is determined. It will be appreciated that the image matching process requires a consistent number and variety of pre-stored images for comparison with the frames of the video file. The comparison process can set a similarity threshold, for example, a 95% similarity threshold is set, so that a frame picture with a similarity of more than 95% with a pre-stored image can represent successful matching, and geographic position information corresponding to the pre-stored image exists in the video file, for example, "eastern bright beads", "white tower park", and the like. Referring to fig. 2, based on the image matching process, it is determined that the video file includes a construction landmark named "eastern pearl", so that the corresponding auxiliary map can be displayed in a floating window in the display area 201 while the video file is played in the display area 101 in fig. 2 according to the geographical location information, and the geographical location information of the "eastern pearl" is expressed in the auxiliary map.
According to the method and the device, the geographic position information contained in the video file can be conveniently and efficiently obtained by matching the frame images in the video file with the pre-stored images, and computing resources are saved.
As an alternative embodiment, step S1 includes:
acquiring a sound file corresponding to a video file;
and identifying the sound file to acquire geographic position information.
In addition to the picture for the video file, the geographical position information may also be obtained by processing a sound file corresponding to the video file. For example, for the format of the video file, a corresponding parsing tool or interface is invoked to obtain the audio track file of the video file. Or the corresponding sound file is obtained by directly picking up the system sound or the play sound of the video file. Preferably, the calibration process is performed on the sound file and the video file in accordance with a time axis or the like. Therefore, geographic position information contained in the sound file can be obtained by analyzing the sound file, and the occurrence time of the geographic position information in the video file and a corresponding picture can be determined based on the matching of the time axes of the sound file and the video file, so that synchronous display can be performed. Specifically, a voice recognition process may be performed for a sound file, and specifically, NLP (Natural Language Processing ) semantic recognition or the like may be employed. For example, geographical location information such as place names in a character in a television series or broadcast content in news is obtained from the character. Even if the sound file without specific text content is detected by means of voiceprint recognition and the like, for example, if the episode in the video file is detected to be "Kramayi song", the corresponding geographic position information is determined to be Kramayi city. Of course, the result of voiceprint detection is not limited to songs, but may be in other forms.
According to the method, the corresponding sound file is obtained based on the video file, and the geographic position information is obtained through identification of the sound file, so that the obtaining dimension of the geographic position information in the video file is enriched, the geographic position information in the video file can be deeply mined, and an auxiliary map with rich content is presented.
As an alternative embodiment, step S1 includes:
acquiring a subtitle file corresponding to a video file;
geographic location information is extracted from the subtitle file.
The subtitle file in this embodiment includes, but is not limited to, formats such as Srt (SubRip Text), lrc (Lyric), smi (subtitle developed by microsoft), and the like, and the text information of the subtitle file may be directly parsed, and the occurrence time of the obtained geographic location information in the video file may be determined based on a time axis. Of course, if the subtitle file cannot be directly acquired, but the video presents subtitle information on the screen by reading the subtitle file, the geographical position information in the subtitle file can be acquired by means of OCR recognition, object detection, image segmentation, or the like for the video screen.
According to the method and the device, the subtitle files in the video files are detected and identified, so that geographic position information contained in the video files can be conveniently and efficiently extracted, the geographic position information is accurately matched with the playing progress of the video files, computing resources are saved, and follow-up synchronous presentation of videos and auxiliary maps is facilitated.
As an alternative embodiment, step S1 includes:
acquiring voice information of a user;
and if the voice information meets the preset requirement, acquiring the geographic position information in the video clip corresponding to the voice information in the video file.
In this embodiment, geographic location information is acquired in response to the intention of a user watching a video, for example, when the video is played to a certain scene, the user sends an exclamation: "where too beautiful-! ", good want to go-! ", or other expression of such semantics; or the user knows well about some place names and speaks an exchange utterance such as "where" and obtains the played video segment according to the origin-destination time matching of the utterance. It is to be understood that the preset requirements for voice information include, but are not limited to, specific phrases, intonation, duration, etc. Preferably, the capturing and detecting operation of the video clip can be triggered by acquiring information such as a barrage meeting the preset requirement and the like input by a user during watching. Preferably, in this embodiment, the processing of the video file, including the processing of the partial video clip, may be sent to the cloud for processing, so as to obtain the geographic location information therein. According to the method and the device, the geographic position information of the video file is detected through voice triggering of the user, so that related operations are more fit with personalized watching requirements of the user.
As an alternative embodiment, step S2 includes:
generating an auxiliary map, and loading map elements corresponding to the geographic position information on the auxiliary map;
the map elements and the time of occurrence of the geographic location information in the video file are displayed in association on the auxiliary map.
The present embodiment generates or calls map elements corresponding to the geographic location information according to the geographic location information. Among other things, map elements typically include map entities presented in points, lines, planes, such as map points of interest (POIs, point of Interest). Step S2, displaying map elements, and displaying the map elements in a reasonable map area according to the number and distribution of the map elements and a proper scale. For example, for a plurality of fishing grounds mentioned in news, the fishing ground closest to the center of the range may be centrally displayed after its entire range is determined. The sitting positions of certain map elements in the whole map range can also be displayed in combination with the eagle eye view, namely the auxiliary map comprises a front view and the eagle eye view. For example, for documentaries related to chinese railways, because the railways are dense, complex and widely distributed, the position of a certain railway mentioned in the video content in the country can be displayed in the eagle eye view of the auxiliary map. In addition, for more dense map elements, a corresponding thermodynamic diagram can be further generated on the basis of scale density level display. For example, for professional videos of the tourist category and the person number of a certain zone, a tourist distribution thermodynamic diagram can be generated according to tourist information in the professional videos. On the basis of the time axis information of the geographic information position in the video file, the auxiliary map can be displayed while the video is played, and on the other hand, the appearance time of the auxiliary map in the video can be displayed on the map element corresponding to the geographic information position, so that the auxiliary map is convenient for a viewer to view. Alternatively, the time for the map element may include, but is not limited to, being expressed by time/minute/second, or in terms of video content, such as "take chinese" third quarter sixth set ". Optionally, a draggable or selectable timeline bar or similar options may be set in the auxiliary map; and highlighting the geographic information position corresponding to the video playing time in a mode of magnifying the map element, highlighting the map element, centrally displaying the map element or a combination thereof and the like according to the time corresponding to the progress bar or the selection item.
According to the method and the device for displaying the map elements, the map elements and the occurrence time of the geographic position information in the video file are displayed on the auxiliary map in an associated mode, so that a user can conveniently, quickly and intuitively find and locate the map elements, and the user can conveniently and quickly and clearly understand and memorize the content of the video file by means of the auxiliary map.
Preferably, after the step of loading the map elements corresponding to the geographic location information on the auxiliary map, the method further comprises:
generating a layer file comprising map elements;
associating and storing the layer file and the video file;
the step of generating an auxiliary map from the geographical location information comprises:
calling a layer file corresponding to the video file;
and generating an auxiliary map according to the layer file.
For the generated map elements, layer files can be built and stored according to the video files or the time periods, the content and the like corresponding to the video files. For example, map elements corresponding to all geographic information positions involved in a movie are saved in a map layer file in Shp (Shape map format supporting ArcGIS) format. Therefore, when the video file is watched later, the auxiliary map can be generated by directly calling the corresponding layer file stored before without detecting and identifying the geographic position information contained in the video file again through the approaches such as image or voice identification. Preferably, the map elements are stored and the display modes thereof are also stored, for example, a display area and a scale in a map, corresponding symbols or special effects, appearance time corresponding to a video file and the like, and of course, the layer file and the map elements therein can be named according to the video file or related information thereof, so that the map elements can be conveniently searched.
Preferably, the stored map elements can be queried, positioned and opened in a state of not playing the video file, the user can collect the interested map elements, and the collection times corresponding to the map elements can reflect the user interest indexes of the corresponding geographic information positions; in addition, the user can generate a corresponding navigation route by opening the layer file and selecting map elements in the layer file as starting points, end points or passing points.
According to the method, the map elements corresponding to the video files are directly associated and stored, so that the auxiliary map can be quickly generated through subsequent calling, and various map services meeting the personalized requirements of users can be realized.
According to the map generation method, the auxiliary map is displayed based on the geographic position information in the video file, and the auxiliary map is synchronously displayed when the video file is played, so that a user can know the geographic position in the video in time in the watching process, a dedicated map of videos such as movies and news is created, personalized watching services are provided for the attention of fans, traveling fans and the like, understanding and impression of the audience on the video are facilitated, and video watching experience is improved.
Example 2
Referring to fig. 3, the present embodiment specifically provides a map generating apparatus, which includes a display unit 1, one or more processing units 2, and a storage unit 3, where the one or more processing units 2 are respectively connected with the storage unit 3 and the display unit 1 in a communication manner;
the storage unit 3 is configured to store instructions that, when executed by the one or more processing units 2, cause the one or more processing units 2 to perform steps comprising:
s51, obtaining geographic position information in a video file;
s52, generating an auxiliary map according to the geographic position information;
s53, synchronously displaying the auxiliary map when the video file is played.
The formats of the video file in step S51 include, but are not limited to, AVI, WMV, MPEG, MP, FLV, ASF, 3GPP, VOB, etc. The geographic position information corresponding to the video file is related information for describing the characteristics of the geographic things. The geographic location information in the video file may be expressed in the form of media of different dimensions, including but not limited to, visual and audio content in the video file. The step S51 of obtaining the geographic location information in the video file may be performed with respect to different media, for example, when the user views the video file in the APP of a certain terminal, the content transferred by the video file may be obtained from different dimensions as long as the APP is authorized to have corresponding rights such as system sound, screen capturing, microphone pickup, etc., and further the geographic location information therein may be obtained by means of OCR, etc.
The auxiliary map in step S52, i.e. the map for assisting the user to watch the video, is usually presented in the form of an electronic map on the same carrier as the video file, but may also be presented in a separate device, a separate screen or in different software. The geographical position information is displayed on the auxiliary map, specifically, the geographical position information can be realized by adding map elements corresponding to the geographical position information in a related map layer, for example, a place name analysis result of 'Sanqingshan' is obtained from a certain video, a corresponding map interest point file is generated, and the map interest point file is loaded in the auxiliary map. Of course, depending on the type of geographic location information, map elements may be presented in dot, line, area, or other forms such as thermodynamic diagrams, for example highlighting a lake or flashing a railway line, etc. The map category includes, but is not limited to, two-dimensional plan, three-dimensional map, indoor map, live-action map, etc., and can also be switched seamlessly according to the type of geographic information location.
Step S53 enables the user to know the geographic location mentioned in the video in time by synchronously displaying the auxiliary map when playing the video file, and the normal viewing progress is not affected, which can be realized based on the display unit 1. The display unit 1 can respectively display the video file and the auxiliary map in the display areas of one or more devices, for example, the display unit plays the video on a first folding screen of a certain folding screen device, and displays the auxiliary map comprising geographic position information on a second folding screen; or respectively playing video files on the screen of a certain mobile phone terminal, and displaying the auxiliary map on the screen of the mobile phone terminal through a transparent and draggable floating window.
The map generation device of the embodiment displays the auxiliary map based on the geographic position information in the video file, and synchronously displays the auxiliary map when the video file is played, so that a user can timely know the geographic position in the video in the watching process, a dedicated map of videos such as movies and news is created, personalized watching services are provided for the interests of fans, travel lovers and the like, understanding and impression of the watched videos are enhanced, and video watching experience is improved
Example 3
Referring to fig. 4, the present embodiment provides an electronic device 30, which includes a processor 31, a memory 32, and a computer program stored in the memory 32 and executable on the processor 31, and the processor 31 implements the map generating method in embodiment 1 when executing the program. The electronic device 30 shown in fig. 4 is merely an example and should not be construed as limiting the functionality and scope of use of embodiments of the present invention.
The electronic device 30 may be in the form of a general purpose computing device, such as a server device. Components of electronic device 30 may include, but are not limited to: the at least one processor 31, the at least one memory 32, a bus 33 connecting the different system components, including the memory 32 and the processor 31. The bus 33 includes a data bus, an address bus, and a control bus.
Memory 32 may include volatile memory such as Random Access Memory (RAM) 321 and/or cache memory 322, and may further include Read Only Memory (ROM) 323.
Memory 32 may also include a program/utility 325 having a set (at least one) of program modules 324, such program modules 324 including, but not limited to: an operating system, one or more application programs, other program modules, and program data, each or some combination of which may include an implementation of a network environment.
The processor 31 executes various functional applications and data processing, such as the map generation method in embodiment 1 of the present invention, by running a computer program stored in the memory 32.
The electronic device 30 may also communicate with one or more external devices 34 (e.g., keyboard, pointing device, etc.). Such communication may be through an input/output (I/O) interface 35. Also, model-generating device 30 may also communicate with one or more networks, such as a Local Area Network (LAN), a Wide Area Network (WAN), and/or a public network, such as the internet, via network adapter 36. Network adapter 36 communicates with the other modules of model-generating device 30 via bus 33. Other hardware and/or software modules may be used in connection with the model-generating device 30, including but not limited to: microcode, device drivers, redundant processors, external disk drive arrays, RAID (disk array) systems, tape drives, data backup storage systems, and the like.
It should be noted that although several units/modules or sub-units/modules of an electronic device are mentioned in the above detailed description, such a division is merely exemplary and not mandatory. Indeed, the features and functionality of two or more units/modules described above may be embodied in one unit/module in accordance with embodiments of the present invention. Conversely, the features and functions of one unit/module described above may be further divided into ones that are embodied by a plurality of units/modules.
Example 4
The present embodiment provides a computer-readable storage medium having stored thereon a computer program which, when executed by a processor, implements the map generation method in embodiment 1.
More specifically, among others, readable storage media may be employed including, but not limited to: portable disk, hard disk, random access memory, read only memory, erasable programmable read only memory, optical storage device, magnetic storage device, or any suitable combination of the foregoing.
In a possible embodiment, the invention may also be implemented in the form of a program product comprising program code for causing a terminal device to carry out the map generation method of embodiment 1 when the program product is run on the terminal device.
Wherein the program code for carrying out the invention may be written in any combination of one or more programming languages, the program code may execute entirely on the user device, partly on the user device, as a stand-alone software package, partly on the user device, partly on a remote device or entirely on the remote device. While specific embodiments of the invention have been described above, it will be appreciated by those skilled in the art that this is by way of example only, and the scope of the invention is defined by the appended claims. Various changes and modifications to these embodiments may be made by those skilled in the art without departing from the principles and spirit of the invention, but such changes and modifications fall within the scope of the invention.

Claims (11)

1. A map generation method, comprising:
obtaining geographic position information in a video file;
generating an auxiliary map according to the geographic position information;
and synchronously displaying the auxiliary map when the video file is played.
2. The map generation method of claim 1, the step of acquiring geographic location information in the video file comprising:
performing target detection processing or image segmentation processing on the video file to obtain text information corresponding to the video file;
and extracting the geographic position information from the text information.
3. The map generation method of claim 1, the step of acquiring geographic location information in the video file comprising:
performing image matching processing on at least one frame of the video file;
and determining the geographic position information according to the image matching processing result.
4. The map generation method of claim 1, the step of acquiring geographic location information in the video file comprising:
acquiring a sound file corresponding to the video file;
and identifying the sound file to acquire the geographic position information.
5. The map generation method of claim 1, the step of acquiring geographic location information in the video file comprising:
acquiring a subtitle file corresponding to the video file;
and extracting the geographic position information from the subtitle file.
6. The map generation method of claim 1, the step of acquiring geographic location information in the video file comprising:
acquiring voice information of a user;
and if the voice information meets the preset requirement, acquiring geographic position information in the video clips corresponding to the voice information in the video file.
7. The map generation method of claim 1, the step of generating an auxiliary map from the geographic location information comprising:
generating the auxiliary map and loading map elements corresponding to the geographic position information on the auxiliary map;
and displaying the appearance time of the map elements and the geographic position information in the video file in an associated mode on the auxiliary map.
8. The map generation method of claim 7, further comprising, after the step of loading the map elements corresponding to the geographic location information by the auxiliary map:
generating a layer file comprising the map elements;
storing the layer file and the video file in a correlated way;
the step of generating an auxiliary map according to the geographic position information comprises the following steps:
calling a layer file corresponding to the video file;
and generating an auxiliary map according to the layer file.
9. A map generating device, characterized in that the map generating device comprises a display unit, one or more processing units and a storage unit, wherein the one or more processing units are respectively connected with the storage unit and the display unit in a communication way;
the storage unit is configured to store instructions that, when executed by the one or more processing units, cause the one or more processing units to perform steps comprising:
obtaining geographic position information in a video file;
generating an auxiliary map according to the geographic position information;
and synchronously displaying the auxiliary map when the display unit plays the video file.
10. An electronic device comprising a memory, a processor and a computer program stored on the memory and executable on the processor, characterized in that the processor implements the map generation method of any of claims 1-9 when the computer program is executed by the processor.
11. A computer-readable storage medium, on which a computer program is stored, characterized in that the computer program, when being executed by a processor, implements the map generation method of any one of claims 1-9.
CN202210520977.0A 2022-05-12 2022-05-12 Map generation method, device, equipment and medium Pending CN117097937A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210520977.0A CN117097937A (en) 2022-05-12 2022-05-12 Map generation method, device, equipment and medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210520977.0A CN117097937A (en) 2022-05-12 2022-05-12 Map generation method, device, equipment and medium

Publications (1)

Publication Number Publication Date
CN117097937A true CN117097937A (en) 2023-11-21

Family

ID=88772281

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210520977.0A Pending CN117097937A (en) 2022-05-12 2022-05-12 Map generation method, device, equipment and medium

Country Status (1)

Country Link
CN (1) CN117097937A (en)

Similar Documents

Publication Publication Date Title
US11849196B2 (en) Automatic data extraction and conversion of video/images/sound information from a slide presentation into an editable notetaking resource with optional overlay of the presenter
US20210056251A1 (en) Automatic Data Extraction and Conversion of Video/Images/Sound Information from a Board-Presented Lecture into an Editable Notetaking Resource
KR101867558B1 (en) Computerized method and device for annotating at least one feature of an image of a view
US9860593B2 (en) Devices, systems, methods, and media for detecting, indexing, and comparing video signals from a video display in a background scene using a camera-enabled device
US20150319510A1 (en) Interactive viewing experiences by detecting on-screen text
US8966372B2 (en) Systems and methods for performing geotagging during video playback
KR101700365B1 (en) Method for providing media-content relation information, device, server, and storage medium thereof
CN109819338A (en) A kind of automatic editing method, apparatus of video and portable terminal
CN106021496A (en) Video search method and video search device
KR102246305B1 (en) Augmented media service providing method, apparatus thereof, and system thereof
US20230071845A1 (en) Interactive viewing experiences by detecting on-screen text
CN102890699A (en) Geotagging of audio recordings
US20110274406A1 (en) Information processing method, information processing device, scene metadata extraction device, loss recovery information generation device, and programs
US9564177B1 (en) Intelligent video navigation techniques
US20160379410A1 (en) Enhanced augmented reality multimedia system
WO2021031733A1 (en) Method for generating video special effect, and terminal
JP6202815B2 (en) Character recognition device, character recognition method, and character recognition program
JP2011040921A (en) Content generator, content generating method, and content generating program
CN113132780A (en) Video synthesis method and device, electronic equipment and readable storage medium
US20190058856A1 (en) Visualizing focus objects from video data on electronic maps
WO2019146466A1 (en) Information processing device, moving-image retrieval method, generation method, and program
CN113992973A (en) Video abstract generation method and device, electronic equipment and storage medium
US10123090B2 (en) Visually representing speech and motion
CN117097937A (en) Map generation method, device, equipment and medium
CN113891026B (en) Recording and broadcasting video marking method and device, medium and electronic equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination