WO2020103447A1 - Procédé et appareil de stockage de type à liaison pour les informations vidéo, dispositif informatique et support d'enregistrement - Google Patents

Procédé et appareil de stockage de type à liaison pour les informations vidéo, dispositif informatique et support d'enregistrement

Info

Publication number
WO2020103447A1
WO2020103447A1 PCT/CN2019/092636 CN2019092636W WO2020103447A1 WO 2020103447 A1 WO2020103447 A1 WO 2020103447A1 CN 2019092636 W CN2019092636 W CN 2019092636W WO 2020103447 A1 WO2020103447 A1 WO 2020103447A1
Authority
WO
WIPO (PCT)
Prior art keywords
information
video
speaker
stored
processed
Prior art date
Application number
PCT/CN2019/092636
Other languages
English (en)
Chinese (zh)
Inventor
吴壮伟
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2020103447A1 publication Critical patent/WO2020103447A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/02Feature extraction for speech recognition; Selection of recognition unit
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8456Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/02Feature extraction for speech recognition; Selection of recognition unit
    • G10L2015/025Phonemes, fenemes or fenones being the recognition units

Definitions

  • This application relates to the field of computer technology, and in particular to a video information chain storage method, device, computer equipment, and storage medium.
  • the traditional processing method is to convert the format of the video file or reduce the resolution of the video file to The compression process is performed to reduce the number of bytes of the video file.
  • the existing video processing method has a problem that the video file cannot be stored lightly.
  • Embodiments of the present application provide a video information chain storage method, device, computer equipment, and storage medium, which are intended to solve the problem in the prior art that the video file cannot be stored lightly.
  • an embodiment of the present application provides a video information chain storage method, which includes: acquiring a video file to be processed, and cutting the video file to be processed through a video cutting model to obtain multiple video segments to be stored; according to a preset
  • the speech recognition model recognizes the voice information in the multiple video segments to be stored to obtain the text information corresponding to the speaker; intercepts the view information corresponding to the text information from the video segment to be stored; according to the text information
  • the speaker stores the obtained text information and view information in a preset database and a linked list corresponding to the speaker.
  • an embodiment of the present application provides a video information chain storage device, which includes a video file cutting unit for acquiring a video file to be processed, and cutting the video file to be processed through a video cutting model to obtain multiple Stored video segment; voice information recognition unit, used to recognize the voice information in the obtained multiple video segments to be stored according to a preset voice recognition model to obtain text information corresponding to the speaker; view information acquisition unit, used To intercept the view information corresponding to the text information from the video segment to be stored; the information storage unit is used to store the obtained text information and view information to the preset database and the linked list corresponding to the speaker according to the speaker corresponding to the text information in.
  • an embodiment of the present application further provides a computer device, including a memory, a processor, and a computer program stored on the memory and executable on the processor, the processor executing the computer
  • the program implements the video information chain storage method described in the first aspect above.
  • an embodiment of the present application further provides a computer-readable storage medium, wherein the computer-readable storage medium stores a computer program, and when the computer program is executed by a processor, the processor causes the processor to execute the On the one hand, the video information chain storage method.
  • FIG. 1 is a schematic flowchart of a video information chain storage method provided by an embodiment of the present application
  • FIG. 2 is a schematic diagram of a sub-process of a video information chain storage method provided by an embodiment of the present application
  • FIG. 3 is a schematic diagram of another sub-process of a video information chain storage method provided by an embodiment of the present application.
  • FIG. 4 is a schematic diagram of another sub-process of a video information chain storage method provided by an embodiment of this application.
  • FIG. 5 is another schematic flowchart of a video information chain storage method provided by an embodiment of the present application.
  • FIG. 6 is a schematic block diagram of a video information chain storage device provided by an embodiment of the present application.
  • FIG. 7 is a schematic block diagram of a subunit of a video information chain storage device provided by an embodiment of this application.
  • FIG. 8 is a schematic block diagram of another subunit of a video information chain storage device provided by an embodiment of the present application.
  • FIG. 9 is a schematic block diagram of another subunit of a video information chain storage device provided by an embodiment of this application.
  • FIG. 10 is another schematic block diagram of a video information chain storage device provided by an embodiment of the present application.
  • FIG. 11 is a schematic block diagram of a computer device provided by an embodiment of the present application.
  • FIG. 1 is a schematic flowchart of a video information chain storage method provided by an embodiment of the present application.
  • the video information chain storage method is applied to terminal devices with information storage functions, such as desktop computers, notebook computers, tablet computers, or mobile phones.
  • the method includes steps S110-S140.
  • S110 Obtain a video file to be processed, and cut the video file to be processed through a video cutting model to obtain multiple video segments to be stored.
  • the user inputs a to-be-processed video file into the user terminal, identifies and cuts the to-be-processed video file through a video cutting model, and obtains multiple video segments to be stored.
  • the to-be-processed video file is a video file input by the user and needs to be stored in a light weight.
  • the to-be-processed video file includes number information, video time stamp, and speaker information.
  • the number information is the number used to identify the to-be-processed video file, that is, the ID of the to-be-processed video file, and each to-be-processed video file has a corresponding number information, and the to-be-processed video file number information is not repeated ;
  • Video timestamp is the information used to mark the time of the video file to be processed. The video timestamp can be used to determine the specific creation time of the video file to be processed;
  • the speaker information is the speaker contained in the video file to be processed Information, one to be processed video file can contain one or more speakers.
  • the corresponding speaker information contains only one speaker; if the to-be-processed video file is a face-to-face interview recording, the corresponding speaker information contains Multiple speakers.
  • step S110 includes sub-steps S111 and S112.
  • the speaker switching time point is obtained through the video cutting model and the speaker information in the to-be-processed video file. Specifically, if the speaker information contains only one speaker, there is no speaker switching time point in the pending video file, and the pending video file is not cut; if the speaker information contains multiple speakers, the If the video file to be processed contains one or more speaker switching time points, the speaker switching time point in the video file to be processed needs to be obtained according to the video cutting model. Recognize the speakers in the video file to be processed through the video cutting model to obtain the switching time point of switching from one speaker to another in the video file to be processed.
  • the video cutting model contains the facial recognition results of all speakers , Face recognition of the speaker in the video file to be processed through the video cutting model can be matched to the speaker of the current picture.
  • Face recognition of the speaker in the video file to be processed through the video cutting model can be matched to the speaker of the current picture.
  • the speech occurs Person switching, obtaining the time when the speaker switch occurs in the two screens of the video file to be processed as the obtained speaker switching time point.
  • S112 Cut the video file to be processed according to the time point when the speaker switches in the video file to be processed to obtain the video segment to be stored corresponding to each speaker.
  • the video file to be processed is cut according to the time point when the speaker switches in the video file to be processed to obtain the video segment to be stored corresponding to each speaker. According to the obtained speaker switching time point, the to-be-processed video file can be cut into multiple video segments to be stored. Specifically, each video segment to be stored corresponds to a speaker, and each video segment to be stored includes its The corresponding time information in the video file to be processed.
  • the voice recognition model is a specific model for recognizing the voice information in the video file. Speech recognition models include acoustic models, speech feature dictionaries, and semantic analysis models.
  • step S120 includes sub-steps S121, S122, and S123.
  • the voice information is segmented according to the acoustic model in the voice recognition model to obtain multiple phonemes contained in the voice information.
  • the voice information is a sentence that the user says by uttering.
  • the voice information received by the user terminal is composed of phonemes that are pronounced by multiple characters, and the phonemes of a character include the frequency and timbre of the pronunciation of the character.
  • the acoustic model contains phonemes for pronunciation of all characters. By matching the phonetic information with all phonemes in the acoustic model, the phonemes of individual characters in the phonetic information can be segmented, and finally the phonetic information contained in the phonetic information can be obtained Multiple phonemes.
  • all phonemes can be converted into pinyin information.
  • the phonetic feature dictionary contains the phoneme information corresponding to the pinyin of all characters.
  • the phoneme of a single character can be converted into the pinyin of the character in the phonetic feature dictionary that matches the phoneme , To convert all phonemes contained in the voice information into pinyin information.
  • the obtained pinyin information is semantically analyzed to convert the pinyin information into text information.
  • the semantic analysis model contains the mapping relationship between the pinyin information and the text information. Through the mapping relationship contained in the semantic analysis model, the obtained pinyin information can be semantically analyzed to convert the pinyin information into text information.
  • the text information corresponding to the pinyin "hé, p ⁇ ng" in the semantic analysis model is "peace”.
  • View information randomly intercepted from the video segment to be stored is view information corresponding to the video segment to be stored. Since each video segment to be stored corresponds to a speaker, the captured view information corresponds to the speaker.
  • the view information may be a video or a picture. By using a video or a picture as the view information corresponding to the video segment to be stored, the view information of the speaker in the stored video segment can be intercepted and saved.
  • the view information is a video
  • the view information is a picture
  • the video segment to be stored Randomly intercept a picture of the speaker as view information corresponding to the video segment to be stored.
  • S140 The obtained text information and view information are stored in the linked list corresponding to the speaker in the preset database according to the speaker corresponding to the text information.
  • the preset database is a database used to store data information.
  • the database contains multiple linked lists.
  • the linked list is a data storage unit that stores text information and view information contained in the video file to be processed according to the time axis.
  • the speaker corresponds to a linked list in the database.
  • the logical order of the data information stored in the linked list is implemented by the pointer linking order in the linked list.
  • the time information corresponding to the text information in the to-be-processed video file is used as the logical order of the linked list, which is
  • the text information and view information in the to-be-processed video file are stored in the linked list by using the time information as the pointer linking order.
  • step S140 includes sub-steps S141, S142, and S143.
  • each video segment to be stored contains its corresponding time information in the video file to be processed, and the text information corresponds to the video segment to be stored in the video file to be processed one by one, therefore, by obtaining the corresponding to the text information to be stored
  • the time information of the video segment in the to-be-processed video file can obtain the corresponding time information of the text information in the to-be-processed video file.
  • time information of a video segment to be stored in the video file to be processed is "1 minute 20 seconds to 3 minutes 10 seconds"
  • “1 minute 20 seconds to 3 minutes 10 seconds” will be used as the The corresponding time information of the corresponding text information in the to-be-processed video file.
  • the text information is stored in the linked list according to the time information of the text information and the corresponding speaker.
  • Each piece of text information corresponds to a speaker.
  • the speaker can obtain a linked list corresponding to the speaker from the preset database, and the text information is stored in the linked list using the time information of the text information as the logical sequence of the linked list, and Add the speaker corresponding to the text message to the stored text message.
  • step S150 is further included after step S140.
  • the index information corresponding to the text information is generated according to the number information of the video file to be processed and the video time stamp and stored in the database.
  • the index information corresponding to the text information can be generated according to the number information and video time stamp of the video file to be processed, and one video file to be processed can correspond to one or more text Information, you need to generate one or more index information correspondingly.
  • certain text information is the third text of the video file to be processed in Table 1, and the corresponding time information of the text information in the video file to be processed is "1 minute 20 seconds to 3 minutes 10 seconds", then the corresponding index is generated The information is "S10021-3, 2018.04.11-1 minutes 20 seconds to 3 minutes 10 seconds".
  • the text information and the view information are saved in the linked list to realize the lightweight storage of the video file to be processed without losing the video
  • the storage space required for the video file is greatly reduced, and very good results have been achieved in the actual application process.
  • An embodiment of the present application further provides a video information chain storage device, which is used to execute any embodiment of the foregoing video information chain storage method.
  • FIG. 6 is a schematic block diagram of a video information chain storage device provided by an embodiment of the present application.
  • the video information chain storage device can be configured in terminal devices such as desktop computers, notebook computers, tablet computers or mobile phones.
  • the video information chain storage device 100 includes a video file cutting unit 110, a voice information recognition unit 120, a view information acquisition unit 130, and an information storage unit 140.
  • the video file cutting unit 110 is used for obtaining a video file to be processed, and cutting the video file to be processed through a video cutting model to obtain multiple video segments to be stored.
  • the video file cutting unit 110 includes subunits: a switching time point acquiring unit 111 and a cutting processing unit 112.
  • the switching time point acquiring unit 111 is configured to obtain the speaker switching time point through the video cutting model and the speaker information in the to-be-processed video file.
  • the cutting processing unit 112 is configured to cut the video file to be processed according to the speaker switching time point in the video file to be processed to obtain the video segment to be stored corresponding to each speaker.
  • the voice information recognition unit 120 is configured to recognize the voice information in the obtained multiple video segments to be stored according to a preset voice recognition model to obtain text information corresponding to the speaker.
  • the voice information recognition unit 120 includes subunits: a phoneme segmentation unit 121, a phonetic information acquisition unit 122 and a text information acquisition unit 123.
  • the phoneme segmentation unit 121 is configured to segment the voice information according to the acoustic model in the voice recognition model to obtain multiple phonemes contained in the voice information.
  • the pinyin information acquisition unit 122 is configured to match the obtained phonemes according to the phonetic feature dictionary in the voice recognition model to convert all phonemes into pinyin information.
  • the text information acquiring unit 123 is configured to perform semantic analysis on the obtained pinyin information according to the semantic analysis model in the speech recognition model to convert the pinyin information into text information.
  • the view information obtaining unit 130 is used to intercept the view information corresponding to the text information from the video segment to be stored.
  • the information storage unit 140 is configured to store the obtained text information and view information in a preset database in a linked list corresponding to the speaker according to the speaker corresponding to the text information.
  • the information storage unit 140 includes subunits: a time information acquisition unit 141, a text information storage unit 142, and a view information storage unit 143.
  • the time information obtaining unit 141 is used to obtain the time information corresponding to the text information in the to-be-processed video file.
  • the text information storage unit 142 is configured to store the text information in a linked list corresponding to the speaker based on the time information of the text information and the corresponding speaker.
  • the view information storage unit 143 is configured to insert the view information corresponding to the text information into the stored text information in the linked list to save the view information.
  • the video information chain storage device 100 further includes a subunit: an index information storage unit 150.
  • the index information storage unit 150 is configured to generate index information corresponding to the text information according to the number information of the video file to be processed and the video time stamp and store it in the database.
  • the above-mentioned video information chain storage device may be implemented in the form of a computer program, and the computer program may run on a computer device as shown in FIG. 11.
  • FIG. 11 is a schematic block diagram of a computer device provided by an embodiment of the present application.
  • the computer device 500 includes a processor 502, a memory, and a network interface 505 connected through a system bus 501, where the memory may include a non-volatile storage medium 503 and an internal memory 504.
  • the non-volatile storage medium 503 can store an operating system 5031 and a computer program 5032.
  • the processor 502 can execute the video information chain storage method.
  • the processor 502 is used to provide computing and control capabilities and support the operation of the entire computer device 500.
  • the internal memory 504 provides an environment for the operation of the computer program 5032 in the non-volatile storage medium 503.
  • the processor 502 can execute the video information chain storage method.
  • the network interface 505 is used for network communication, such as the transmission of data information.
  • the network interface 505 is used for network communication, such as the transmission of data information.
  • FIG. 11 is only a block diagram of a part of the structure related to the solution of the present application, and does not constitute a limitation on the computer device 500 to which the solution of the present application is applied.
  • the specific computer device 500 may include more or less components than shown in the figure, or combine certain components, or have a different arrangement of components.
  • the processor 502 is used to run the computer program 5032 stored in the memory to implement the video information chain storage method of the present application.
  • the embodiment of the computer device shown in FIG. 11 does not constitute a limitation on the specific configuration of the computer device.
  • the computer device may include more or fewer components than shown in the figure. Or combine certain components, or arrange different components.
  • the computer device may only include a memory and a processor. In such an embodiment, the structures and functions of the memory and the processor are consistent with the embodiment shown in FIG. 11, and details are not described herein again.
  • the processor 502 may be a central processing unit (Central Processing Unit, CPU), and the processor 502 may also be other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), Application specific integrated circuit (Application Specific Integrated Circuit, ASIC), ready-made programmable gate array (Field-Programmable Gate Array, FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc.
  • the general-purpose processor may be a microprocessor or the processor may be any conventional processor.
  • a computer-readable storage medium may be a non-volatile computer-readable storage medium.
  • the computer-readable storage medium stores a computer program, where the computer program is executed by the processor to implement the video information chain storage method of the embodiments of the present application.
  • the storage medium may be an internal storage unit of the foregoing device, such as a hard disk or a memory of the device.
  • the storage medium may also be an external storage device of the device, such as a plug-in hard disk equipped on the device, a smart memory card (Smart) Card (SMC), a secure digital (SD) card, or a flash memory card (Flash Card) etc.
  • the storage medium may also include both an internal storage unit of the device and an external storage device.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Signal Processing (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

La présente invention concerne un procédé et un appareil de stockage de type à liaison pour les informations vidéo, un dispositif informatique et un support d'enregistrement. Le procédé consiste à : obtenir un fichier vidéo à traiter, et segmenter ledit fichier vidéo au moyen d'un modèle de segmentation vidéo pour obtenir une pluralité de segments vidéo à stocker ; reconnaître, selon un modèle de reconnaissance vocale prédéfini, des informations vocales dans la pluralité obtenue de segments vidéo, pour obtenir des informations textuelles correspondant au locuteur ; écrêter, à partir desdits segments vidéo, des informations de visualisation correspondant aux informations de texte ; stocker, en fonction du locuteur correspondant aux informations de texte, les informations de texte et les informations de visualisation obtenues dans une liste liée correspondant au locuteur dans une base de données prédéfinie.
PCT/CN2019/092636 2018-11-21 2019-06-25 Procédé et appareil de stockage de type à liaison pour les informations vidéo, dispositif informatique et support d'enregistrement WO2020103447A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201811389154.9A CN109582823A (zh) 2018-11-21 2018-11-21 视频信息链式存储方法、装置、计算机设备及存储介质
CN201811389154.9 2018-11-21

Publications (1)

Publication Number Publication Date
WO2020103447A1 true WO2020103447A1 (fr) 2020-05-28

Family

ID=65923631

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/092636 WO2020103447A1 (fr) 2018-11-21 2019-06-25 Procédé et appareil de stockage de type à liaison pour les informations vidéo, dispositif informatique et support d'enregistrement

Country Status (2)

Country Link
CN (1) CN109582823A (fr)
WO (1) WO2020103447A1 (fr)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109582823A (zh) * 2018-11-21 2019-04-05 平安科技(深圳)有限公司 视频信息链式存储方法、装置、计算机设备及存储介质
CN110414352B (zh) * 2019-06-26 2022-11-29 深圳职业技术学院 从视频文件中提取ppt文件信息的方法及相关设备
CN114697706A (zh) * 2022-03-29 2022-07-01 深圳市恒扬数据股份有限公司 一种视频内容处理方法、装置、终端及存储介质
CN115129198B (zh) * 2022-06-13 2023-10-27 中移互联网有限公司 一种数据获取方法、装置、电子设备及存储介质

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110205330A1 (en) * 2010-02-25 2011-08-25 Ricoh Company, Ltd. Video conference system, processing method used in the same, and machine-readable medium
CN105512348A (zh) * 2016-01-28 2016-04-20 北京旷视科技有限公司 用于处理视频和相关音频的方法和装置及检索方法和装置
CN105957531A (zh) * 2016-04-25 2016-09-21 上海交通大学 基于云平台的演讲内容提取方法及装置
CN106162222A (zh) * 2015-04-22 2016-11-23 无锡天脉聚源传媒科技有限公司 一种视频镜头切分的方法及装置
CN107241616A (zh) * 2017-06-09 2017-10-10 腾讯科技(深圳)有限公司 视频台词提取方法、装置及存储介质
CN109582823A (zh) * 2018-11-21 2019-04-05 平安科技(深圳)有限公司 视频信息链式存储方法、装置、计算机设备及存储介质

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104301649B (zh) * 2014-10-29 2019-08-16 上海斐讯数据通信技术有限公司 一种视频拍摄存储、播放的方法和系统
CN105912615A (zh) * 2016-04-05 2016-08-31 重庆大学 一种基于人类语音内容索引的音频和视频文件管理方法
CN107273423B (zh) * 2017-05-15 2019-04-12 中国移动通信集团湖北有限公司 多媒体消息数据处理方法、装置和系统
CN108829765A (zh) * 2018-05-29 2018-11-16 平安科技(深圳)有限公司 一种信息查询方法、装置、计算机设备及存储介质

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110205330A1 (en) * 2010-02-25 2011-08-25 Ricoh Company, Ltd. Video conference system, processing method used in the same, and machine-readable medium
CN106162222A (zh) * 2015-04-22 2016-11-23 无锡天脉聚源传媒科技有限公司 一种视频镜头切分的方法及装置
CN105512348A (zh) * 2016-01-28 2016-04-20 北京旷视科技有限公司 用于处理视频和相关音频的方法和装置及检索方法和装置
CN105957531A (zh) * 2016-04-25 2016-09-21 上海交通大学 基于云平台的演讲内容提取方法及装置
CN107241616A (zh) * 2017-06-09 2017-10-10 腾讯科技(深圳)有限公司 视频台词提取方法、装置及存储介质
CN109582823A (zh) * 2018-11-21 2019-04-05 平安科技(深圳)有限公司 视频信息链式存储方法、装置、计算机设备及存储介质

Also Published As

Publication number Publication date
CN109582823A (zh) 2019-04-05

Similar Documents

Publication Publication Date Title
WO2020103447A1 (fr) Procédé et appareil de stockage de type à liaison pour les informations vidéo, dispositif informatique et support d'enregistrement
US10079014B2 (en) Name recognition system
WO2019095586A1 (fr) Procédé de génération de comptes rendus de réunions, serveur d'application et support d'informations lisible par ordinateur
WO2020087655A1 (fr) Procédé, appareil et dispositif de traduction, et support de stockage lisible
WO2019232991A1 (fr) Procédé de reconnaissance de voix de conférence sous forme de texte, dispositif électronique et support de stockage
US11138971B2 (en) Using context to interpret natural language speech recognition commands
US10811005B2 (en) Adapting voice input processing based on voice input characteristics
US10270736B2 (en) Account adding method, terminal, server, and computer storage medium
US11011170B2 (en) Speech processing method and device
JP5860171B2 (ja) 入力処理方法及び装置
CN110164435A (zh) 语音识别方法、装置、设备及计算机可读存储介质
US10108698B2 (en) Common data repository for improving transactional efficiencies of user interactions with a computing device
WO2019223134A1 (fr) Procédé et appareil de recherche de message vocal, dispositif informatique et support d'enregistrement
WO2016023317A1 (fr) Procédé et terminal de traitement d'informations vocales
WO2020119064A1 (fr) Procédé et dispositif de stockage d'informations internet de façon liée, appareil informatique, et support d'enregistrement
WO2021135603A1 (fr) Procédé de reconnaissance d'intention, serveur et support de stockage
WO2013163804A1 (fr) Procédé et dispositif de réglage d'une banque de mots
TW201606750A (zh) 使用外國字文法的語音辨識
WO2021218069A1 (fr) Procédé et appareil de traitement interactif basé sur une configuration de scénario dynamique, et dispositif informatique
CN108682421B (zh) 一种语音识别方法、终端设备及计算机可读存储介质
CN106713111B (zh) 一种添加好友的处理方法、终端及服务器
WO2020037921A1 (fr) Procédé et appareil d'invites d'images d'expression, dispositif informatique, et support de stockage
TW201339862A (zh) 語言歧義消除系統及方法
US8868419B2 (en) Generalizing text content summary from speech content
KR20190074508A (ko) 챗봇을 위한 대화 모델의 데이터 크라우드소싱 방법

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19887972

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19887972

Country of ref document: EP

Kind code of ref document: A1