CN108764007A - Based on OCR with text analysis technique to the measurement method of attention - Google Patents
Based on OCR with text analysis technique to the measurement method of attention Download PDFInfo
- Publication number
- CN108764007A CN108764007A CN201810138534.9A CN201810138534A CN108764007A CN 108764007 A CN108764007 A CN 108764007A CN 201810138534 A CN201810138534 A CN 201810138534A CN 108764007 A CN108764007 A CN 108764007A
- Authority
- CN
- China
- Prior art keywords
- attention
- time
- ocr
- text
- picture
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/40—Document-oriented image-based pattern recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/279—Recognition of textual entities
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/30—Semantic analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/22—Image preprocessing by selection of a specific region containing or referencing a pattern; Locating or processing of specific regions to guide the detection or recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Artificial Intelligence (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- General Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
The invention discloses the measurement methods based on OCR and text analysis technique to attention, pass through the behavioral data of the technical limit spacings person of being observed such as OCR, the concept that the person's of being observed attention is mapped in semantic space is calculated with the text analysis techniques such as such as feature selecting, similarity measurement, measures distribution and the transfer method of attention indirectly.Traditional psychology is absorbed in the power for measuring this cognitive ability of attention with neural brain science, and the signified attention of the present invention belongs to management science scope, refer to people pay close attention to a theme, event lasting scale.The method of the present invention has great use in personal management, business administration, commending system etc..
Description
Technical field
The invention belongs to attention fields of measurement, it is directed generally to attention distribution and transfer, tool in research semantic space
Body is related to measuring attention distribution and the transfer in semantic space based on OCR and text analysis technique.
Background technology
What is attention, it is to be noted that human brain realizes the process occupied by external or inbeing.And attention refers to
The ability that Mr. Yu plants things is directed toward and is concentrated to the psychological activity of people.The founder Michael H.Goldhaber of attention economy are carried
It arrives, today's society is the abundant society even spread unchecked of an information maximum, relative to superfluous information, the attention resource of people
It is rare, with the development of information age, valuable not instead of information, attention.People increasingly focus on the time
Management, it is desirable to limited attention is placed in more significant things, continued more in the market about the book of time management
The portable computer software of fast sale, all kinds of time managements emerges one after another, and has great meaning in the information age to the measurement of attention
Justice.
In recent years, scientists develop the measuring technique of attention from many aspects.On the one hand it is to paying attention to energy
Power, for example, attention measuring technique is considered that a kind of Cognitive Aptitude Test, the method for Test of attention have note by psychological professional
Meaning power test chart tests table, test exercise etc.;In addition, may be preferably by eyeball itself together with head to the measurement of attention
The movement in portion is evaluated, and when your eyes stare at certain, your prodigious probability of attention just at this, more as a result, may be used
It is referred to as attention monitor, such as eye tracker to monitor head and oculomotor wearable device;Attention neuro-physiology
Family is fed back by studying the brain wave of brain wave variation and human body, is developed brain dateline helmet and is calculated attention.On the other hand, expand
Notice that the time scale of process, space scale enter hand measurement note from a series of theme of the concerns generated by attention process, event
Meaning power.For example, attention behavior of Netease's cloud music software by recorder to music, forms unique commending system;
Rescue TimeAPP meeting counting users are judged that user has spent respectively and how long opened using the time of App according to classification
Hair, design, chat, amusement and other above, user according to statistics can obtain distribution of the attention on APP.Ihour is one
Money user's typing task, logger task execute the time management software of duration, the thing record that oneself is actively absorbed in by user
On ihour, it is desirable to obtain oneself input degree to everything feelings.The attention of these two aspects measures, and is mankind itself's
Attention training provides a great help with self-management constraint.
Nowadays, big and small electronic equipment can all install a screen, be designed to obtain human attention,
And the recovery time of the mankind present 80% or more, all it is to see staring at various screens.Exactly, the content in screen
Obtain the attention of people.Based on such situation, the present invention is proposed to be measured based on the personal attention of OCR and text analysis technique
Method, from larger time and space scale, in conjunction with above-mentioned both sides measurement angle, to existing attention measurement side
Formula is proposed technical perfect and is improved, and is obtained expression of the theme of the person's of being observed concern in semantic space, is finally obtained reality
When, the distribution of long-term attention, flow regime.
Baidupedia is the world of language meaning to the definition of semantic space.In general, information is meaning and symbol
Entity, inherent meaning only pass through certain external form (symbols such as action, expression, word, speech, picture, image)
It can just express.Therefore, each symbolism is all the language for conveying meaning, the meaning structure expressed by them in a broad sense
At specific semantic space.Accordingly, the present invention extracts the intrinsic meaning for paying close attention to theme from the angle of external attention
Come, variation of the attention on semantic space is showed by the meaning.
OCR technique (Optical Character Recognition), i.e. optical character identification refer to electronic equipment inspection
The character printed on paper is looked into, its shape is determined by the pattern for detecting dark, bright, is then translated into shape with character identifying method
The process of computword.The input of OCR identifying systems is the image of different-format, is exported as computword, purpose is very
Simply, it is that image is converted, makes the figure in image continue to preserve, has text of the table then in table in data and image
Word becomes computword without exception, and the enabled storage capacity reduction for reaching image data, the word identified can reuse and divide
Analysis, can also save the manpower inputted by keyboard and time.Current OCR technique comparative maturity, for the recognition accuracy of certificate
99% can generally be reached.The Text region accuracy of image quality clearly normal image also can reach 85% or more.Personal vision
The information such as the computer webpage of involved browsing, mobile phone screen, using OCR technique, can easily be converted into computer can
The word of identification extracts expression of the theme in semantic space of the person's of being observed concern.
Text analysis technique, using natural language processing (NLP:Natural Language Processing) and analysis
Content of text is converted into data by method, establishes its mathematical model, scientific abstraction is carried out to text, to describe and replace
Text.It enables a computer to realize the identification to text by calculating to this model and operation.Text analyzing generally by
Three steps form, and parse data, search retrieval, text mining.Text analyzing at present be widely used in customer experience, customer insight,
Data analysis etc..From semantic space angle, using text analysis technique, using semantic information as data set, in conjunction with
Temporal information, the attention that can obtain each period are mapped in concept in semantic space, obtain attention dwell point with
And the information such as flowing of attention, attention is indirectly measured to realize.
Invention content
The present invention starts with from external attention and the angle of semantic space, and daily Fixed Time Interval (second/minute) is to being seen
Electronic equipment operated by survey person intercepts frame individual's image browsing and converts image to analyzable text using OCR technique
Collection.Semantic information is extracted from rambling text set using text analysis technique, analysis attention is bonded with the time
Semantic data collection.Meanwhile the present invention will record the number and time that each subprogram is activated, configuration program data
Collection can be obtained from the property of subprogram in relation to personal attention properties macroscopical on spatio-temporal distribution.Two
A data set friendship is mutually echoed, and reflects the distribution characteristics of attention.For example, the sub- journey that everyone activates in some specific time
Arrange in order number distribution, learning program and chat program Automobile driving etc..From data set, this method can obtain, attention
In the mapping of semantic space, the transfer of attention dwell point, attention at any time.According to the personal attention structure generated daily
Feature can accumulate the long-term observation to be formed to personal attention feature, to analyze personal attention knot at work
More long-term attention features such as structure.
The present invention proposes to include the following steps the measurement method of attention based on OCR and text analysis technique:
Step 1. captures visual range
1-1) visual range captured has significant change with the variation for intercepting frequency, and herein, we to locate daily
It intercepts within 8 hours in working condition, for fixed every 1 minute interception one action picture;
The code of interception personal work picture 1-2) is write, adds up 60 frame pictures of capture per hour, daily accumulative capture 480
Frame work picture.
Step 2.OCR technical finesses
The capturing visual handled using OCR technique identifies the word in picture, generates text set;
Step 3. obtains data set
3-1) text set is segmented, removes the processing such as noise generation text set;
3-2) using the participle collection of removal noise, a feature extraction is done to the text of each picture, is extracted crucial semantic
Information records time and semantic information, generative semantics data set;
3-3) obtain the same day activation process of electronic equipment and activationary time and the shut-in time of process, configuration program number
According to collection.
Step 4. calculates attention distribution characteristics
It will 4-1) activate program corresponding with label, and calculate the subprogram label distribution activated in certain specific time and number
Distribution;
It 4-2) obtains attention and shifts flow network, record semanteme is at the time of shift and the keyword at the moment, root
According to the sequencing of time, generates attention and shift flow network;
Attention residence time 4-3) is calculated, when the time is with where node B where attention shifts the node A in flow network
Between subtract each other, the residence time of node A can be obtained;Under long-term observation, the situation of change of long-term dwell point in time can be formed;
Focus 4-4) is calculated, the focus of each period daily is recorded, generates the focus curve on the same day, further
, in the case where long-time is observed, form long-term focus curve.
Advantageous effect
1, compared to use the methods of eye tracker, psychology test pay attention to cognitive ability.This method mainly measures more wide
General attention, measurement is the theme of concern, things distribution in time and transfer case by paying attention to generating.This measurement
Method does not cause radiation to body, is not impacted to personal work, do not limit individual's without dressing heavy mechanical equipment
Scope of activities, observation are convenient;
2, compared to Rescue Time, the time managements software such as Ihour, what this method measured is the pass by paying attention to generating
Mapping of the theme, things of note on semantic space, rather than the things itself paid close attention to, what is covered can reflect attention feature
Information is more comprehensive, without being manually entered, can realize and automatically record.
3, long-term monitoring is conveniently formed to the Attention behavior for the person of being observed, and skill is measured compared to traditional attention
Art can observe the attention change that some short-term attentional power measurements do not observe.
Description of the drawings
Fig. 1 is attention method flow schematic diagram of the present invention;
Fig. 2 is DBOW model schematics.
Specific implementation mode
Technical scheme of the present invention is described in detail below in conjunction with the accompanying drawings:
The thinking of the present invention is to collect the electronic equipment use information of object being observed, specifically includes program and uses interface
Information, handle use information using statistical method and OCR technique, these information indirects illustrate object being observed attention
Mapping of the information on semantic space finally obtains the distribution characteristics of attention according to these information.
The basic procedure of the method for the present invention is as shown in Figure 1, specifically include following steps:
Step 1. captures visual range
Module Pillow, Time, selenium timing (for example, one minute) interception that image is handled using Python is seen
The currently used electronic curtain of survey person;Daily observation period is working time, the morning 9:00—12:00, afternoon 13:00-
17:00, amount to 8 hours, per hour 60 frame pictures of accumulative capture, daily 480 frames of accumulative capture work picture, by all pictures
It stores to same file under pressing from both sides.Fig. 2 is the work picture example of frame interception.
Step 2.OCR technical finesses
Darg screen is stored, ApiOCR is the API service that Baidu's AI Text regions provide, API
(Application Programming Interface, application programming interface) is that some are pre-defined
Function, it is therefore an objective to application program be provided and be able to access the ability of one group of routine based on certain software or hardware with developer.We
The use of ApiOCR is one by one text by the On-Screen Identification of interception.Word generation table 2 corresponding with the time in the figure of identification, by table 2
Information stored into lexicographic text set in Python.
Table 1 generates text set
Step 3. obtains data set
3-1) the NTLK modules of python is used to carry out subordinate sentence processing first to text set, then carry out word segmentation processing, simultaneously
Filter out useless information.Such as, count one day within participle word frequency, remove some only minority work pictures in occur it is low
Frequency word;Remove the stop words such as the auxiliary word for not carrying any information, conjunction, except being text set after making an uproar;Stop words is illustrated:
{about、above、according、accordingly、across、actually、after、afterwards、again、
against,ain't,all};
3-2) the attention object of human brain of the information response in text set, namely meaning object is accounted for, extraction accounts for the semantic letter of meaning object
Breath generative semantics data set is a feature selecting (Feature using the participle collection after denoising to the text of each picture
Selection).Text feature selection, refer to extracting from original feature (all texts) it is a small amount of, it is representative
Feature, but the type of feature does not change, and is originally text lexical set, is still vocabulary after feature extraction, but quantity is big
It is big to reduce.The keyword that participle is concentrated is extracted by text feature selection, analyzes the attention of worker.Use TF-IDF spies
Selection algorithm is levied, is as follows:
TF-IDF=word frequency (TF) × inverse document frequency (IDF) (3)
The value for calculating the TF-IDF of each word in document, then arranges according to descending, takes several words of front as special
Levy attribute.Here due to before only taking K it is big, the concept of semantic space is mapped in as attention.
Where 3-3) keyword that feature extraction goes out is considered as the attention of author per minute, we term it attentions
Point similarly counts the participle collection of daily 480 pictures, and using feature extraction keyword, these keywords are attention in language
The expression of justice spatially;
3-4) (for example, 1 minute) calls process on using windows PowerShell at every fixed time
API obtain existing program process, process computer memory occupancy situation.According to time, journey of the statistics computer in observation time
Computer memory occupancy situation when the time span of sort run, distinct program operation, the program number of different time operation.
Step 4. calculates attention distribution characteristics
It 4-1) is concentrated from program data, activation program is corresponding with program tag database, and the class that such as works, is learned amusement class
It is corresponding to practise class, chat class etc., calculates on one day time shaft, certain special time period activates the label distribution of program.It calculates
Subprogram type and the number distribution activated in certain specific time;
Table 2 activates program tag database
Program | Label |
Python | Programming, work |
Wechat (wechat) | Chat |
R | Programming, work |
It seeks survival danger spot | Game |
Tencent's video | Amusement |
Word | Work |
…… | …… |
4-2) common property gives birth to 480 participle collection daily, text similarity analysis is done using Doc2vec, if two texts are similar
It spends low, it is believed that attention has occurred transfer and generates attention transfer flow network within this time.Record lime light shifts
At the time of and the keyword at the moment generate attention and shift flow network according to the sequencing of time.The node of network is
Lime light, the oriented even side between node A and node B, looks like for the attention force A of last moment, subsequent time is transferred to
Pay attention to force B;Doc2vec algorithm principles are as follows, obtain
The vector of Sentence/Document indicates that also there are two types of models by Doc2Vec, respectively:Distributed
Memory (DM) and Distributed Bag ofWords (DBOW), DM models given context and document to
Predict that the probability of word, DBOW models predict one group of random word in document in the case of given document vector in the case of amount
Probability.Here the present invention uses DBOW models, and the input of the model is the vector of document, and prediction is taken out at random in the document
The word of sample
Attention residence time 4-3) is calculated, time and -1 institute of node i where attention shifts the node i in flow network
Subtract each other in the time, the residence time of node i -1 can be obtained;Under normal conditions, we calculate residence time formula it is as follows,
Focus=top10 { max (time (nodei)-time(nodei-1)) i ∈ (1,2,3 ... n) (4)
Under long-term observation, the situation of change of long-term dwell point in time can be formed;
Focus 4-4) is calculated, the calculation formula of focus is:
(1-n/N) × 100% (5)
Wherein n is the number of network node for each period including, and N is the participle collection number generated each period.For example,
The focus of each hour is calculated, N is 60 at this time, if morning 9:00-10:00 produces 10 nodes, then focus is 90%.
The focus of record daily each period, generates the focus curve on the same day, further, in the case where long-time is observed, is formed and is grown
The focus curve of phase.
Claims (4)
1. based on OCR with text analysis technique to the measurement method of attention, which is characterized in that include the following steps:
Step 1. captures visual range:
Step 2.OCR technical finesses:
The capturing visual handled using OCR technique identifies the word in picture, generates text set;
Step 3. obtains data set:
Step 4. calculates attention distribution characteristics.
2. the method as described in claim 1, which is characterized in that the step 1 includes:
1-1) visual range captured has significant change with the variation for intercepting frequency;With in running order daily 8
Hour interception, fixed every 1 minute interception one action picture;
The code of interception personal work picture 1-2) is write, adds up 60 frame pictures of capture per hour, daily 480 frame works of accumulative capture
Make picture.
3. the method as described in claim 1, which is characterized in that the step 2 includes:
3-1) text set is segmented, removes the processing such as noise generation text set;
3-2) using the participle collection of removal noise, a feature extraction is done to the text of each picture, extracts crucial semantic information,
Record time and semantic information, generative semantics data set;
3-3) obtain the same day activation process of electronic equipment and activationary time and the shut-in time of process, configuration program data
Collection.
4. the method as described in claim 1, which is characterized in that the step 4 includes:
It will 4-1) activate program corresponding with label, and calculate the subprogram label distribution activated in certain specific time and number point
Cloth;
4-2) obtain attention and shift flow network, record semanteme is at the time of shift and the keyword at the moment, according to when
Between sequencing, generate attention shift flow network;
Attention residence time 4-3) is calculated, time where attention shifts the node A in flow network and time phase where node B
Subtract, obtains the residence time of node A;Under long-term observation, the situation of change of long-term dwell point in time is formed;
Focus 4-4) is calculated, the focus of each period daily is recorded, generates the focus curve on the same day, further,
Under observing for a long time, long-term focus curve is formed.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810138534.9A CN108764007A (en) | 2018-02-10 | 2018-02-10 | Based on OCR with text analysis technique to the measurement method of attention |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810138534.9A CN108764007A (en) | 2018-02-10 | 2018-02-10 | Based on OCR with text analysis technique to the measurement method of attention |
Publications (1)
Publication Number | Publication Date |
---|---|
CN108764007A true CN108764007A (en) | 2018-11-06 |
Family
ID=63980065
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810138534.9A Pending CN108764007A (en) | 2018-02-10 | 2018-02-10 | Based on OCR with text analysis technique to the measurement method of attention |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN108764007A (en) |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109522953A (en) * | 2018-11-13 | 2019-03-26 | 北京师范大学 | The method classified based on internet startup disk algorithm and CNN to graph structure data |
CN109800434A (en) * | 2019-01-25 | 2019-05-24 | 陕西师范大学 | Abstract text header generation method based on eye movement attention |
CN111276149A (en) * | 2020-01-19 | 2020-06-12 | 科大讯飞股份有限公司 | Voice recognition method, device, equipment and readable storage medium |
CN111506196A (en) * | 2020-04-21 | 2020-08-07 | 合肥凯石投资咨询有限公司 | Pupil screen compounding method for attention evaluation |
US20210209356A1 (en) * | 2020-01-06 | 2021-07-08 | Samsung Electronics Co., Ltd. | Method for keyword extraction and electronic device implementing the same |
CN113343981A (en) * | 2021-06-16 | 2021-09-03 | 北京百度网讯科技有限公司 | Visual feature enhanced character recognition method, device and equipment |
CN115100664A (en) * | 2022-06-20 | 2022-09-23 | 济南大学 | Multi-mode false news identification method and system based on correlation information expansion |
US11956400B2 (en) | 2022-08-30 | 2024-04-09 | Capital One Services, Llc | Systems and methods for measuring document legibility |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101918939A (en) * | 2008-01-15 | 2010-12-15 | 亚马逊技术有限公司 | Enhancing and storing data for recall and use |
CN102893327A (en) * | 2010-03-19 | 2013-01-23 | 数字标记公司 | Intuitive computing methods and systems |
US20140253701A1 (en) * | 2013-03-10 | 2014-09-11 | Orcam Technologies Ltd. | Apparatus and method for analyzing images |
CN106164959A (en) * | 2014-02-06 | 2016-11-23 | 威图数据研究公司 | Behavior affair system and correlation technique |
CN106537290A (en) * | 2014-05-09 | 2017-03-22 | 谷歌公司 | Systems and methods for biomechanically-based eye signals for interacting with real and virtual objects |
CN106708261A (en) * | 2016-12-05 | 2017-05-24 | 深圳大学 | Brain-computer interaction-based attention training method and system |
CN107169049A (en) * | 2017-04-25 | 2017-09-15 | 腾讯科技(深圳)有限公司 | The label information generation method and device of application |
CN107193803A (en) * | 2017-05-26 | 2017-09-22 | 北京东方科诺科技发展有限公司 | A kind of particular task text key word extracting method based on semanteme |
-
2018
- 2018-02-10 CN CN201810138534.9A patent/CN108764007A/en active Pending
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101918939A (en) * | 2008-01-15 | 2010-12-15 | 亚马逊技术有限公司 | Enhancing and storing data for recall and use |
CN102893327A (en) * | 2010-03-19 | 2013-01-23 | 数字标记公司 | Intuitive computing methods and systems |
US20140253701A1 (en) * | 2013-03-10 | 2014-09-11 | Orcam Technologies Ltd. | Apparatus and method for analyzing images |
CN106164959A (en) * | 2014-02-06 | 2016-11-23 | 威图数据研究公司 | Behavior affair system and correlation technique |
CN106537290A (en) * | 2014-05-09 | 2017-03-22 | 谷歌公司 | Systems and methods for biomechanically-based eye signals for interacting with real and virtual objects |
CN106708261A (en) * | 2016-12-05 | 2017-05-24 | 深圳大学 | Brain-computer interaction-based attention training method and system |
CN107169049A (en) * | 2017-04-25 | 2017-09-15 | 腾讯科技(深圳)有限公司 | The label information generation method and device of application |
CN107193803A (en) * | 2017-05-26 | 2017-09-22 | 北京东方科诺科技发展有限公司 | A kind of particular task text key word extracting method based on semanteme |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109522953A (en) * | 2018-11-13 | 2019-03-26 | 北京师范大学 | The method classified based on internet startup disk algorithm and CNN to graph structure data |
CN109800434A (en) * | 2019-01-25 | 2019-05-24 | 陕西师范大学 | Abstract text header generation method based on eye movement attention |
CN109800434B (en) * | 2019-01-25 | 2023-07-18 | 陕西师范大学 | Method for generating abstract text title based on eye movement attention |
US20210209356A1 (en) * | 2020-01-06 | 2021-07-08 | Samsung Electronics Co., Ltd. | Method for keyword extraction and electronic device implementing the same |
CN111276149A (en) * | 2020-01-19 | 2020-06-12 | 科大讯飞股份有限公司 | Voice recognition method, device, equipment and readable storage medium |
CN111276149B (en) * | 2020-01-19 | 2023-04-18 | 科大讯飞股份有限公司 | Voice recognition method, device, equipment and readable storage medium |
CN111506196A (en) * | 2020-04-21 | 2020-08-07 | 合肥凯石投资咨询有限公司 | Pupil screen compounding method for attention evaluation |
CN113343981A (en) * | 2021-06-16 | 2021-09-03 | 北京百度网讯科技有限公司 | Visual feature enhanced character recognition method, device and equipment |
CN115100664A (en) * | 2022-06-20 | 2022-09-23 | 济南大学 | Multi-mode false news identification method and system based on correlation information expansion |
CN115100664B (en) * | 2022-06-20 | 2024-04-09 | 济南大学 | Multi-mode false news identification method and system based on correlation information expansion |
US11956400B2 (en) | 2022-08-30 | 2024-04-09 | Capital One Services, Llc | Systems and methods for measuring document legibility |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108764007A (en) | Based on OCR with text analysis technique to the measurement method of attention | |
KR101830061B1 (en) | Identifying activities using a hybrid user-activity model | |
Pohl et al. | Online indexing and clustering of social media data for emergency management | |
CN106683688B (en) | Emotion detection method and device | |
AU2011358100B2 (en) | Systems and methods for image-to-text and text-to-image association | |
Arapakis et al. | A comparison of general vs personalised affective models for the prediction of topical relevance | |
CN106537387B (en) | Retrieval/storage image associated with event | |
Chen et al. | In the eyes of the beholder: analyzing social media use of neutral and controversial terms for COVID-19 | |
Alexandridis et al. | A knowledge-based deep learning architecture for aspect-based sentiment analysis | |
Kedar et al. | Automatic personality assessment: A systematic review | |
Yadav et al. | Review on automated depression detection from audio visual clue using sentiment analysis | |
Jindal et al. | Intend to analyze Social Media feeds to detect behavioral trends of individuals to proactively act against Social Threats | |
Serbaya | [Retracted] Analyzing the Role of Emotional Intelligence on the Performance of Small and Medium Enterprises (SMEs) Using AI‐Based Convolutional Neural Networks (CNNs) | |
Constantine et al. | A framework for emotion recognition from human computer interaction in natural setting | |
Dudzik et al. | Exploring personal memories and video content as context for facial behavior in predictions of video-induced emotions | |
Bi et al. | Detecting depression on sina microblog using depressing domain lexicon | |
CN112668384B (en) | Knowledge graph construction method, system, electronic equipment and storage medium | |
Bayram | Revealing the reflections of the pandemic by investigating COVID-19 related news articles using machine learning and network analysis | |
Walawalkar et al. | Support vector learning for gender classification using audio and visual cues: A comparison | |
Burzo et al. | Sensing affective response to visual narratives | |
Pan et al. | Automatically infer human traits and behavior from social media data | |
KR100330333B1 (en) | System and method , computer program and its recording media for an informational diagnosis of personal character | |
Chen et al. | A web-based longitudinal mental health monitoring system | |
Perner | Novel Methods for Forensic Multimedia Data Analysis: Part I | |
Kotenko et al. | Analysis of Formats of Young People’s Communicative Behavior in Social Network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
DD01 | Delivery of document by public notice | ||
DD01 | Delivery of document by public notice |
Addressee: Ji Zhi Academy (Beijing) science and Technology Co., Ltd. Document name: Notification to Make Rectification Addressee: Ji Zhi Academy (Beijing) science and Technology Co., Ltd. Document name: Notice of non patent agent (person) |
|
DD01 | Delivery of document by public notice | ||
DD01 | Delivery of document by public notice |
Addressee: Ji Zhi Xue Yuan (Beijing) Technology Co., Ltd. (patent owner) Document name: Notification of Passing Examination on Formalities |
|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WD01 | Invention patent application deemed withdrawn after publication | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20181106 |