CN104462454A - Character analyzing method - Google Patents
Character analyzing method Download PDFInfo
- Publication number
- CN104462454A CN104462454A CN201410780141.XA CN201410780141A CN104462454A CN 104462454 A CN104462454 A CN 104462454A CN 201410780141 A CN201410780141 A CN 201410780141A CN 104462454 A CN104462454 A CN 104462454A
- Authority
- CN
- China
- Prior art keywords
- video
- audio
- data
- audio frequency
- video information
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G06F19/34—
Landscapes
- Health & Medical Sciences (AREA)
- Engineering & Computer Science (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
The invention provides a character analyzing method which is realized on the basis of a personal mobile terminal. The character analyzing method includes acquiring audio-video information in the personal mobile terminal through a data acquiring module, performing incomplete data filtering and redundant data filtering on the audio-video information acquired by the data acquiring module by a data filtering and analyzing module; using data after being filtered to establish an audio-video information database; comparing text similarity between the audio-video information database acquired by the data filtering and analyzing module and preset character attributes and a content corpus by a character analyzing module to judge character type to which a user of the personal mobile terminal belongs. The character analyzing method has the advantages that data volume is targeted, data security is guaranteed, data accuracy is higher and analysis result is more accurate.
Description
Technical field
The present invention relates to a kind of character analysis method.
Background technology
Children and adolescent period, the personality of a people moulds most important period just, busy due to aspects such as present crowd's work, causes, and the communication exchange with children is also fewer, causes understanding few to the following personality trend of children.And the complicacy of social phenomenon and human communication now, children are caused more easily to occur psychological problems, current children's character analysis mainly relies on Mental Health Counseling and teaches mechanism, and major part makes auxiliary property treatment behavior again when children's psychology personality defectiveness, lacks preventative.And some is made character analysis by data such as networks and also there is a little defect, network data is originated wide Analysis of Complex, and whether data are that my network behavior at ordinary times exists dispute, the accuracy of data is not high, and some data relates to individual privacy, network data is also easily acquired when transmitting, and safety has hidden danger.Also some analyzes personality (as judged user mood and personality according to call voice) by cellphone information, and data are here all call voices at that time, and the personality now analyzing out is relevant with user's mood at that time, not too accurately.
Summary of the invention
The invention provides a kind of character analysis method, data volume is targeted, data security is secure, data accuracy is higher, analysis result is more accurate.
In order to achieve the above object, the invention provides a kind of character analysis method, the method realizes based on individual mobile terminal, comprises following steps:
Step S1, data acquisition module obtain the audio/video information in individual mobile terminal;
Step S2, data filtering parsing module carry out deficiency of data filtration to the audio/video information that data acquisition module obtains and redundant data is filtered, and sets up audio/video information database by the data after filtering;
The audio/video information database that data filtering analysis module obtains by step S3, character analysis module and preset personality attribute and corpus of content carry out text similarity comparison, judge which kind of character type the user of individual mobile terminal belongs to.
In described step S1, data acquisition module obtains audio/video information from audio player, video player and webpage.
In described step S1, data acquisition module obtains audio-frequency information from audio player, and described audio-frequency information comprises audio frequency name, audio presentations person, album name, reproduction time and broadcasting time.
In described step S1, data acquisition module obtains video information from video player, and described video information comprises video name, reproduction time and broadcasting time.
In described step S1, data acquisition module obtains audio/video information from webpage, and described audio/video information comprises network address, browsing time, audio frequency and video name, audio frequency and video separator, audio presentations person/video presentations person, number of visits, whether plays mark and whether download mark.
When data acquisition module obtains audio/video information from webpage, if the audio frequency and video directly obtained are by name empty, then use web page contents analytic technique, the text message on webpage is resolved, from text message, obtains audio frequency and video name.
In described step S2, data filtering parsing module carries out deficiency of data bag filter to the audio frequency and video name that data acquisition module obtains and contains: carry out text similarity to audio frequency and video name and compare, if text similarity is more than or equal to threshold value, then thinks identical audio frequency and video.
In described step S2, data filtering parsing module carries out redundant data bag filter to the audio frequency and video name that data acquisition module obtains and contains: be called basis for estimation with audio frequency and video, removes and repeats record.
In described step S2, the audio/video information in described audio/video information database comprises audio frequency and video name, audio presentations person/video presentations person, audio frequency and video separator, reproduction time and broadcasting time.
Described personality attribute and corpus of content comprise audio frequency and video name, audio presentations person/video presentations person, audio frequency and video separator, audio frequency and video type and the character type corresponding with audio frequency and video type and personality content.
In described step S3, described character analysis module is called critical field with audio frequency and video, in the audio/video information database obtained by data filtering analysis module, broadcasting time exceedes and plays the audio/video information data of threshold value and preset personality attribute and corpus of content and carry out text similarity comparison, if the similarity of audio frequency and video name is more than or equal to compare threshold, then judge that the user of individual mobile terminal belongs to the character type of this audio frequency and video name correspondence.
The invention has the advantages that:
1, data source of the present invention gathers mainly for the amusement behavior of children's every day use mobile terminal, and data volume is targeted.
2, data of the present invention come from the native data of children's every day use mobile terminal, and without the need to through Internet Transmission, data security is secure.
3, the data collected are carried out data filtering parsing and character analysis carries out text similarity comparison by the present invention, make data accuracy higher.
4, personality attribute of the present invention and corpus of content, adopts autonomous learning to add rich language material database data, makes analysis result more accurate.
Accompanying drawing explanation
Fig. 1 is process flow diagram of the present invention.
Fig. 2 is the particular flow sheet in embodiments of the invention.
Embodiment
Illustrate preferred embodiment of the present invention according to Fig. 1 and Fig. 2 below.
As depicted in figs. 1 and 2, the invention provides a kind of character analysis method, the method realizes based on individual mobile terminal, comprises following steps:
Step S1, data acquisition module obtain the audio/video information in individual mobile terminal.
Described audio/video information derives from audio player, video player and webpage.
Described step S1 comprises following steps:
Step S1.1, data acquisition module obtain audio-frequency information from audio player, and described audio-frequency information comprises audio frequency name, audio presentations person, album name, reproduction time and broadcasting time, and above-mentioned audio-frequency information is inserted table 1.
Table 1
Step S1.2, data acquisition module obtain video information from video player, and described video information comprises video name, reproduction time and broadcasting time, and above-mentioned video information is inserted table 2.
Table 2
Step S1.3, data acquisition module obtain audio/video information from webpage, described audio/video information comprises network address, browsing time, audio frequency and video name, audio frequency and video separator, audio presentations person/video presentations person, number of visits, whether plays mark and whether download mark, and above-mentioned audio/video information is inserted table 3.
Table 3
In the present embodiment, described audio frequency and video separator, with " 0 " mark audio frequency, uses one token video; Described whether broadcasting mark, marks with " 0 " and does not play, play by one token; Described whether download mark, marks with " 0 " and does not download, download by one token.
Data acquisition module is when obtaining audio/video information from webpage, because when browsing or play some website, these webpages not necessarily provide complete audio frequency and video name, at this moment just need to use web page contents analytic technique, text message on webpage is resolved, from text message, obtains audio frequency and video name.
Therefore, in step S1.3, data acquisition module, when obtaining the audio frequency and video name in audio/video information, carries out step S1.4, judges whether from webpage, directly to obtain audio frequency and video name, namely, judge whether audio frequency and video name is empty, if so, then carries out step S1.5, if not, then directly audio frequency and video name is obtained.
Step S1.5, data acquisition module are resolved the text message on webpage, obtain audio frequency and video name from text message.
Data acquisition module utilization web page contents analytic technique carrys out the label on filtering web page, and finds crucial label to obtain text message, and crucial label here refers to similar song title, singer, types of songs, film, film types etc.
Step S2, data filtering parsing module carry out deficiency of data filtration to the audio frequency and video name that data acquisition module obtains and redundant data is filtered, and sets up audio/video information database by the data after filtering.
Described step S2 comprises following steps:
Step S2.1, data filtering parsing module carry out deficiency of data filtration to the audio frequency and video name that data acquisition module obtains and redundant data is filtered.
Described deficiency of data bag filter contains: carry out text similarity to audio frequency and video name and compare, if text similarity is more than or equal to threshold value, then thinks identical audio frequency and video.
In the present embodiment, threshold value is 80%, uses participle technique that audio frequency and video name is divided into several key word, then carries out text similarity to these key words and compare.
Described redundant data bag filter contains: be called basis for estimation with audio frequency and video, removes and repeats record.
Step S2.2, data filtering parsing module set up audio/video information database.
Audio/video information in described audio/video information database comprises audio frequency and video name, audio presentations person/video presentations person, audio frequency and video separator, reproduction time and broadcasting time.
Described audio/video information database is as shown in table 4.
Table 4
In the present embodiment, described audio frequency and video separator, with " 0 " mark audio frequency, uses one token video.
The audio/video information database that data filtering analysis module obtains by step S3, character analysis module and preset personality attribute and corpus of content carry out text similarity comparison, judge which kind of character type the user of individual mobile terminal belongs to.
As shown in chart 5, described personality attribute and corpus of content comprise audio frequency and video name, audio presentations person/video presentations person, audio frequency and video separator, audio frequency and video type and the character type corresponding with audio frequency and video type and personality content.
Table 5
In the present embodiment, described audio frequency and video separator, with " 0 " mark audio frequency, uses one token video; Described audio frequency and video type, as audio frequency belong to popular, miss old times or old friends, sentiment, rock and roll etc.; Video belongs to ancient costume, violence, emotion, science fiction etc.; Described character type, as " sanguine temperament ", " quality of bile ", " lymphatic temperament " and " melancholy " four kinds of personality; Described personality content is the detailed explanation to character type.
Described character analysis module is called critical field with audio frequency and video, in the audio/video information database obtained by data filtering analysis module, broadcasting time exceedes and plays the audio/video information data of threshold value and preset personality attribute and corpus of content and carry out text similarity comparison, if the similarity of audio frequency and video name is more than or equal to compare threshold, then judge that the user of individual mobile terminal belongs to the character type of this audio frequency and video name correspondence.
In the present embodiment, described broadcasting threshold value is 1, and described compare threshold is 80%, uses participle technique that audio frequency and video name and audio presentations person/video presentations person are divided into several key word, then carries out text similarity to these key words and compare.
The autonomous learning that character analysis method provided by the invention also comprises step S4, personality attribute and corpus of content is perfect.
According to the result of character analysis, personality attribute and corpus of content add the corresponding relation of audio/video information data and character type further, make the data grows of personality attribute and corpus of content comprehensive, analyze the personality drawn more accurate.
The present invention, by analyzing the personality information obtained, is presented to user, is convenient to the personality grasping the user using individual mobile terminal, especially uses the personality of the children of individual mobile terminal, thus the healthy benign development of correct guiding children.The present invention not only can the personality whether defectiveness of Timeliness coverage children, and can analyze the following personality of children, and these personality had both embodied negative side also can embody positive one side; More crucially go out the personality of children according to these data accurate analysis, grasp in time the personality trend of children, whether defectiveness etc., and accomplish to prevent in advance.
Although content of the present invention has done detailed introduction by above preferred embodiment, will be appreciated that above-mentioned description should not be considered to limitation of the present invention.After those skilled in the art have read foregoing, for multiple amendment of the present invention and substitute will be all apparent.Therefore, protection scope of the present invention should be limited to the appended claims.
Claims (11)
1. a character analysis method, is characterized in that, the method realizes based on individual mobile terminal, comprises following steps:
Step S1, data acquisition module obtain the audio/video information in individual mobile terminal;
Step S2, data filtering parsing module carry out deficiency of data filtration to the audio/video information that data acquisition module obtains and redundant data is filtered, and sets up audio/video information database by the data after filtering;
The audio/video information database that data filtering analysis module obtains by step S3, character analysis module and preset personality attribute and corpus of content carry out text similarity comparison, judge which kind of character type the user of individual mobile terminal belongs to.
2. character analysis method as claimed in claim 1, it is characterized in that, in described step S1, data acquisition module obtains audio/video information from audio player, video player and webpage.
3. character analysis method as claimed in claim 2, it is characterized in that, in described step S1, data acquisition module obtains audio-frequency information from audio player, and described audio-frequency information comprises audio frequency name, audio presentations person, album name, reproduction time and broadcasting time.
4. character analysis method as claimed in claim 2, it is characterized in that, in described step S1, data acquisition module obtains video information from video player, and described video information comprises video name, reproduction time and broadcasting time.
5. character analysis method as claimed in claim 2, it is characterized in that, in described step S1, data acquisition module obtains audio/video information from webpage, and described audio/video information comprises network address, browsing time, audio frequency and video name, audio frequency and video separator, audio presentations person/video presentations person, number of visits, whether plays mark and whether download mark.
6. character analysis method as claimed in claim 5, it is characterized in that, when data acquisition module obtains audio/video information from webpage, if the audio frequency and video directly obtained are by name empty, then use web page contents analytic technique, text message on webpage is resolved, from text message, obtains audio frequency and video name.
7. as the character analysis method in claim 1-6 as described in any one, it is characterized in that, in described step S2, data filtering parsing module carries out deficiency of data bag filter to the audio frequency and video name that data acquisition module obtains and contains: carry out text similarity to audio frequency and video name and compare, if text similarity is more than or equal to threshold value, then think identical audio frequency and video.
8. character analysis method as claimed in claim 7, it is characterized in that, in described step S2, data filtering parsing module carries out redundant data bag filter to the audio frequency and video name that data acquisition module obtains and contains: be called basis for estimation with audio frequency and video, removes and repeats record.
9. character analysis method as claimed in claim 8, it is characterized in that, in described step S2, the audio/video information in described audio/video information database comprises audio frequency and video name, audio presentations person/video presentations person, audio frequency and video separator, reproduction time and broadcasting time.
10. character analysis method as claimed in claim 9, it is characterized in that, described personality attribute and corpus of content comprise audio frequency and video name, audio presentations person/video presentations person, audio frequency and video separator, audio frequency and video type and the character type corresponding with audio frequency and video type and personality content.
11. character analysis methods as claimed in claim 10, it is characterized in that, in described step S3, described character analysis module is called critical field with audio frequency and video, in the audio/video information database obtained by data filtering analysis module, broadcasting time exceedes and plays the audio/video information data of threshold value and preset personality attribute and corpus of content and carry out text similarity comparison, if the similarity of audio frequency and video name is more than or equal to compare threshold, then judge that the user of individual mobile terminal belongs to the character type of this audio frequency and video name correspondence.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410780141.XA CN104462454A (en) | 2014-12-17 | 2014-12-17 | Character analyzing method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410780141.XA CN104462454A (en) | 2014-12-17 | 2014-12-17 | Character analyzing method |
Publications (1)
Publication Number | Publication Date |
---|---|
CN104462454A true CN104462454A (en) | 2015-03-25 |
Family
ID=52908489
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201410780141.XA Pending CN104462454A (en) | 2014-12-17 | 2014-12-17 | Character analyzing method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN104462454A (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108021864A (en) * | 2017-11-02 | 2018-05-11 | 平安科技(深圳)有限公司 | Character personality analysis method, device and storage medium |
CN108038414A (en) * | 2017-11-02 | 2018-05-15 | 平安科技(深圳)有限公司 | Character personality analysis method, device and storage medium based on Recognition with Recurrent Neural Network |
CN109146430A (en) * | 2018-09-05 | 2019-01-04 | 福建省伯乐仁资智能科技有限公司 | A kind of Online Video interview method and system |
CN111297379A (en) * | 2020-02-10 | 2020-06-19 | 中国科学院深圳先进技术研究院 | Brain-computer combination system and method based on sensory transmission |
WO2021159230A1 (en) * | 2020-02-10 | 2021-08-19 | 中国科学院深圳先进技术研究院 | Brain-computer interface system and method based on sensory transmission |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102254265A (en) * | 2010-05-18 | 2011-11-23 | 北京首家通信技术有限公司 | Rich media internet advertisement content matching and effect evaluation method |
CN102541590A (en) * | 2011-12-14 | 2012-07-04 | 奇智软件(北京)有限公司 | Software recommending method and system |
CN103294795A (en) * | 2013-05-24 | 2013-09-11 | 华东师范大学 | Method for adjusting film recommending diversity by utilizing users' characters |
CN104091610A (en) * | 2013-11-15 | 2014-10-08 | 腾讯科技(深圳)有限公司 | Audio file managing method and device |
CN104202620A (en) * | 2014-07-04 | 2014-12-10 | 南京超聚通信科技有限公司 | System and method of implementing video on demand and review of unidirectional set-top box through mobile intelligent terminal |
-
2014
- 2014-12-17 CN CN201410780141.XA patent/CN104462454A/en active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102254265A (en) * | 2010-05-18 | 2011-11-23 | 北京首家通信技术有限公司 | Rich media internet advertisement content matching and effect evaluation method |
CN102541590A (en) * | 2011-12-14 | 2012-07-04 | 奇智软件(北京)有限公司 | Software recommending method and system |
CN103294795A (en) * | 2013-05-24 | 2013-09-11 | 华东师范大学 | Method for adjusting film recommending diversity by utilizing users' characters |
CN104091610A (en) * | 2013-11-15 | 2014-10-08 | 腾讯科技(深圳)有限公司 | Audio file managing method and device |
CN104202620A (en) * | 2014-07-04 | 2014-12-10 | 南京超聚通信科技有限公司 | System and method of implementing video on demand and review of unidirectional set-top box through mobile intelligent terminal |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108021864A (en) * | 2017-11-02 | 2018-05-11 | 平安科技(深圳)有限公司 | Character personality analysis method, device and storage medium |
CN108038414A (en) * | 2017-11-02 | 2018-05-15 | 平安科技(深圳)有限公司 | Character personality analysis method, device and storage medium based on Recognition with Recurrent Neural Network |
WO2019085330A1 (en) * | 2017-11-02 | 2019-05-09 | 平安科技(深圳)有限公司 | Personal character analysis method, device, and storage medium |
WO2019085329A1 (en) * | 2017-11-02 | 2019-05-09 | 平安科技(深圳)有限公司 | Recurrent neural network-based personal character analysis method, device, and storage medium |
CN109146430A (en) * | 2018-09-05 | 2019-01-04 | 福建省伯乐仁资智能科技有限公司 | A kind of Online Video interview method and system |
CN111297379A (en) * | 2020-02-10 | 2020-06-19 | 中国科学院深圳先进技术研究院 | Brain-computer combination system and method based on sensory transmission |
WO2021159230A1 (en) * | 2020-02-10 | 2021-08-19 | 中国科学院深圳先进技术研究院 | Brain-computer interface system and method based on sensory transmission |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN106953887B (en) | Fine-grained radio station audio content personalized organization recommendation method | |
Pozzi et al. | Challenges of sentiment analysis in social networks: an overview | |
Torgersen et al. | A corpus-based study of pragmatic markers in London English | |
CN104462454A (en) | Character analyzing method | |
KR101605430B1 (en) | SYSTEM AND METHOD FOR BUINDING QAs DATABASE AND SEARCH SYSTEM AND METHOD USING THE SAME | |
CN104915443B (en) | A kind of abstracting method of Chinese microblogging evaluation object | |
CN106250553A (en) | A kind of service recommendation method and terminal | |
CN107885745A (en) | A kind of song recommendations method and device | |
EP3087505A1 (en) | System and methods for vocal commenting on selected web pages | |
CN104731874B (en) | A kind of evaluation information generation method and device | |
Nozza et al. | A multi-view sentiment corpus | |
Hong | The power of Bollywood: A study on opportunities, challenges, and audiences’ perceptions of Indian cinema in China | |
Snell | Social class and language | |
Raamkumar et al. | Understanding the Twitter usage of humanities and social sciences academic journals | |
KR101326313B1 (en) | Method of classifying emotion from multi sentence using context information | |
CN106934049B (en) | News question selection analysis method and device | |
Lacasse et al. | # Yoga on instagram: Understanding the nature of yoga in the online conversation and community | |
CN108170845A (en) | Multimedia data processing method, device and storage medium | |
KR20170034481A (en) | Intelligent mobile recommended music service system and method | |
CN108416015A (en) | A kind of information security method for pushing | |
CN110019921B (en) | Audio and attribute association method and device and audio searching method and device | |
Panwar et al. | Impact of fake news on readers’ usage behaviour for news items on facebook and twitter | |
Wu et al. | When diversity meets speciality: Friend recommendation in online social networks | |
Ahmad et al. | Personality prediction of Malaysian Facebook users: cultural preferences and features variation | |
Carter | Going Gaga: Pop fandom as online community of practice |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20150325 |
|
RJ01 | Rejection of invention patent application after publication |