CN114723009B - Data representation method and system based on asynchronous event stream - Google Patents

Data representation method and system based on asynchronous event stream Download PDF

Info

Publication number
CN114723009B
CN114723009B CN202210379153.6A CN202210379153A CN114723009B CN 114723009 B CN114723009 B CN 114723009B CN 202210379153 A CN202210379153 A CN 202210379153A CN 114723009 B CN114723009 B CN 114723009B
Authority
CN
China
Prior art keywords
event
doe
moe
time
polarity
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202210379153.6A
Other languages
Chinese (zh)
Other versions
CN114723009A (en
Inventor
古富强
郭方明
陈超
黄柳金
刘淑文
张栋宇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chongqing University
Original Assignee
Chongqing University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chongqing University filed Critical Chongqing University
Priority to CN202210379153.6A priority Critical patent/CN114723009B/en
Publication of CN114723009A publication Critical patent/CN114723009A/en
Application granted granted Critical
Publication of CN114723009B publication Critical patent/CN114723009B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/049Temporal neural networks, e.g. delay elements, oscillating neurons or pulsed inputs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/06Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
    • G06N3/063Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Abstract

The invention relates to a data representation method and system based on an asynchronous event stream, and belongs to the technical field of deep learning feature engineering. The method obtains richer representation of asynchronous event data by splicing event size (MOE) information and event Density (DOE) information, and comprises the following steps: s1, inputting an asynchronous event stream as event data; s2, acquiring event size MOE; s3, acquiring an event density DOE; s4, splicing the MOE and the DOE to obtain MDOE expression, wherein the obtained MDOE simultaneously comprises size information of event data and density information of the event. Compared with the existing event-based representation method, the MDOE method provided by the invention contains more abundant information about the polarity, time and density of the event, and has two advantages: 1) It is a conceptual simple generic representation, task independent. 2) The present method achieves superior performance relative to existing representations on various event-based datasets.

Description

Data representation method and system based on asynchronous event stream
Technical Field
The invention belongs to the technical field of deep learning feature engineering, and relates to a data representation method and system based on asynchronous event streams.
Background
Asynchronous event streams refer to output event streams through event-based sensors (DVS cameras) having a higher dynamic range, higher event resolution, lower time delay, and better energy efficiency than conventional devices (RGB cameras), and are widely used in fields such as autopilot, safety monitoring, and industrial automation. However, due to the nature of the asynchronous event stream itself, learning from these data presents many challenges, which cannot be directly used for training of convolutional neural networks.
Learning a useful representation that can be exploited by convolutional neural networks through the use of asynchronous event streams is one approach to solving this challenge. Currently, the asynchronous event stream representation mode mainly comprises an event representation through artificial design, a pulse-based event representation, a learning-based event representation and the like. A naive idea of the artificial-based event representation approach is to use the number of events that occur per pixel, but this way event information and polarity information of the events are discarded, which may degrade the performance of downstream applications. There are many improvements in this respect, such as taking polarity information into account, introducing an event surface concept to describe the recent history of event spatial neighborhood events, but this approach is susceptible to noise events due to the relatively expensive computation of the time surface, and only taking into account the timestamp of the last event in the vicinity of the event. The impulse-based representation mainly utilizes impulse neural networks, which are used for many event-based tasks, but their usability in real-world scenarios is limited because impulse neural network neurons are not microscopic. There are ways to train the neural network with frame-based data and convert the learned data to a pulsed neural network and to approximate the pulse derivatives with a similar function, but the accuracy achieved by the pulse-based approach is still lower than the deep learning approach to date. There are many learning-based representation methods currently handling asynchronous event data, such as a grid-based representation method called EST, which can learn end-to-end features directly from asynchronous event-based data by microkernel convolution and quantization, but which need to be adapted according to downstream tasks. In the Matrix LSTM approach, a grid of LSTM cells is used to learn an end-to-end event-based representation that can capture local temporal features while preserving spatial structure. PhaseLSTM is a variant of the LSTM unit in which the network uses time gates to learn the exact timing of incoming events, extracting relevant features through the word embedding layer, but does not capture generic features well due to the limited representation capabilities of the embedding layer. Current presentation methods have the problem of discarding one or more types of information regarding event polarity, density, or time information, which may lead to downstream program performance degradation.
Therefore, it is necessary to focus on how to efficiently represent the event stream data-based method to expand the application of the event data-oriented learning model.
Disclosure of Invention
In view of the above, the present invention is directed to a method and a system for representing data based on an asynchronous event stream, which can obtain a richer representation of asynchronous event data by concatenating event size (MOE) information and event Density (DOE) information, so as to solve the problem that one or more types of information in event polarity, density or time information are lost in the existing method for representing data based on an asynchronous event stream, and avoid the performance degradation of downstream programs.
In order to achieve the above purpose, the present invention provides the following technical solutions:
a method of data representation based on an asynchronous event stream, the method comprising the steps of: s1, inputting an asynchronous event stream as event data; s2, acquiring event size MOE; s3, acquiring an event density DOE; s4, splicing the MOE and the DOE to obtain MDOE expression, wherein the obtained MDOE simultaneously comprises size information of event data and density information of the event.
Further, in step S2, the sampling method is to set a fixed time interval when MOE is acquired, and sample at each fixed time interval to obtain a representation with dimensions of 2×c×h×w, where H and W are the spatial height and width of the event image data, respectively, and C is the number of time bins; the method is similar to the neural dynamics in a pulsed neuron, except that in a pulsed neuron, when the membrane potential exceeds the threshold of the neuron, the pulse is fired while the membrane potential is reset, whereas in this step the MOE acquisition has no firing process of the membrane potential and therefore the membrane potential is not reset, thus preserving all polarity information.
Further, in step S3, the event count over the whole event period is not directly considered in the process of acquiring the event density DOE, but the event count over each time window is used to generate a vector of size C for each pixel, and by doing this for each polarity, a DOE representation of size 2 xcxhxw is obtained.
Further, in step S1, let ε be an asynchronous event input sequence expressed as:
Figure BDA0003591545460000021
wherein x is i Is the position (x for DVS camera i =(x i ,y i ) Coordinates of the pixel that triggered the event), t i Is the timestamp of event generation, p i Is the polarity of the event, which takes two values: 1 and-1, respectively representing positive and negative events; i is the event number and the dynamics for impulse neurons are described as follows:
Figure BDA0003591545460000022
where u (t) represents the intimal potential of the neuron at time t,
Figure BDA0003591545460000023
is a weighted sum of the pre-neuron inputs, τ is a time constant, and over a period of time, the denser the events, the greater the membrane potential.
Further, in step S2 and step S3, the MOE and DOE are obtained by sampling the asynchronous event sequence, and it is assumed that E represents the MOE and V represents the DOE, which is described as follows:
Figure BDA0003591545460000031
Figure BDA0003591545460000032
k(a)=max(0,1-|a|) (5)
Figure BDA0003591545460000033
Figure BDA0003591545460000034
t n =t 1 +(c n +1)ΔT (8)
wherein ε is + And epsilon-is a sequence of events having positive and negative polarity, respectively; k (a) is a bilinear kernel that mimics neuronal dynamics, 1 ei Is an indicator function that ensures that only events between specified time periods are counted; (x) l ,y m ,c n ) Is the space-time coordinates, x, on the voxel grid l ∈{0,1···,W-1},y m E {0, 1. H-1}, and c n E {0, 1. N-1}; Δt is the size of the time bin; to reduce the effect of sensor noise (e.g., different event numbers reporting accurate scenes at different times), the resulting MOE (E ± ) And DOE (V) ± ) Divided by their respective maximum values over the pixel and time bins, normalized.
Further, in step S4, the MDOE is a splice of the MOE and DOE along the polarity, denoted by S, i.e
S(p,x l ,y m ,c n )=Cat ± (E ± (x l ,y m ,c n ),V ± (x l ,y m ,c n )) (9)
Where Cat denotes the splice operation, p denotes the polarity of the MOE and DOE, p e {0,1, 2, 3}, where p=0 and p=1 denote the on-off polarity of the MOE, and p=2 and p=3 correspond to the on-off polarity of the DOE.
The invention has the beneficial effects that:
compared with the existing event-based representation method, the MDOE method provided by the invention contains more abundant information about the polarity, time and density of the event, and has two advantages: 1) It is a conceptual simple generic representation, task independent. 2) The present method achieves superior performance relative to existing representations on various event-based datasets.
Additional advantages, objects, and features of the invention will be set forth in part in the description which follows and in part will become apparent to those having ordinary skill in the art upon examination of the following or may be learned from practice of the invention. The objects and other advantages of the invention may be realized and obtained by means of the instrumentalities and combinations particularly pointed out in the specification.
Drawings
For the purpose of making the objects, technical solutions and advantages of the present invention more apparent, the present invention will be described in the following preferred detail with reference to the accompanying drawings, in which:
FIG. 1 is a schematic diagram of membrane potential changes of impulse neurons and a schematic diagram of MOE membrane potential changes of the same sequence;
FIG. 2 is a pseudo code schematic diagram of the proposed method of the present invention;
FIG. 3 is a schematic diagram of the time delay impact assessment on N-Caltech 101;
fig. 4 is a schematic diagram of the effect of using training data of different scales.
Detailed Description
The technical scheme of the invention is described in detail below with reference to the accompanying drawings.
Fig. 1 is a schematic diagram of membrane potential change of a pulse neuron and a schematic diagram of MOE membrane potential change of the same sequence, fig. 2 is a pseudo code schematic diagram of the method according to the present invention, fig. 3 is a schematic diagram of time delay influence evaluation on N-Caltech101, and fig. 4 is a schematic diagram of effect of training data using different proportions.
The data representation method based on the asynchronous event stream provided by the invention comprises the following steps: s1, inputting an asynchronous event stream as event data; s2, acquiring event size MOE; s3, acquiring an event density DOE; s4, splicing the MOE and the DOE to obtain MDOE expression, wherein the obtained MDOE simultaneously comprises size information of event data and density information of the event.
In step S2, the sampling method is to set a fixed time interval when MOE is acquired, and sample at each fixed time interval to obtain a representation with dimensions of 2 xcxhxw, where H and W are the spatial height and width of the event image data, respectively, and C is the number of time bins; the method is similar to the neural dynamics in impulse neurons, except that in impulse neurons when the membrane potential exceeds the threshold of the neuron, the impulse is triggered and the membrane potential is reset, whereas in this step the MOE acquisition has no triggering of the membrane potential and therefore the membrane potential is not reset, thus preserving all polarity information, as shown in fig. 1, fig. 1 (a) shows the dynamic change of the neuron membrane potential and fig. 1 (b) shows the MOE of the same event sequence.
In step S3, the event count over the whole event period is not directly taken into account in the process of obtaining the event density DOE, but rather the event count over each time window is used to generate a vector of size C for each pixel, and by doing this for each polarity, a DOE representation of size 2 xcxhxw is obtained.
As shown in fig. 2, let epsilon be an asynchronous event input sequence, expressed as:
Figure BDA0003591545460000041
wherein x is i Is the position (x for DVS camera i =(x i ,y i ) Coordinates of the pixel that triggered the event), t i Is the timestamp of event generation, p i Is the polarity of the event, which takes two values: 1 and-1, respectively representing positive and negative events; i is the event number and the dynamics for impulse neurons are described as follows:
Figure BDA0003591545460000042
where u (t) represents the intimal potential of the neuron at time t,
Figure BDA0003591545460000051
is a weighted sum of the pre-neuron inputs, τ is a time constant, and over a period of time, the denser the events, the greater the membrane potential.
The MOE and DOE are obtained separately for asynchronous event sequence samples, assuming E represents MOE and V represents DOE, as follows:
Figure BDA0003591545460000052
Figure BDA0003591545460000053
k(a)=max(0,1-|a|) (5)
Figure BDA0003591545460000054
Figure BDA0003591545460000055
t n =t 1 +(c n +1)ΔT (8)
wherein ε is + And epsilon - A sequence of events having positive and negative polarity, respectively; k (a) is a bilinear kernel that mimics neuronal dynamics, 1 ei Is an indicator function that ensures that only events between specified time periods are counted; (x) l ,y m ,c n ) Is the space-time coordinates, x, on the voxel grid l ∈{0,1···,W-1},y m E {0, 1. H-1}, and c n E {0, 1. N-1}; Δt is the size of the time bin; to reduce the effect of sensor noise (e.g., different event numbers reporting accurate scenes at different times), the resulting MOE (E ± ) And DOE (V) ± ) Divided by their respective maximum values over the pixel and time bins, normalized.
MDOE is a splice of MOE and DOE along polarity, denoted by S, i.e
S(p,x l ,y m ,c n )=Cat ± (E ± (x l ,y m ,c n ),V ± (x l ,y m ,c n )) (9)
Where Cat denotes the splice operation, p denotes the polarity of the MOE and DOE, p e {0,1, 2, 3}, where p=0 and p=1 denote the on-off polarity of the MOE, and p=2 and p=3 correspond to the on-off polarity of the DOE.
The new asynchronous event data representation method provided by the invention not only comprises the polarity and time information of the event, but also comprises the event density, and can better represent the information of asynchronous event stream data, in the embodiment, the method provided by the invention is evaluated through three disclosed event-based data sets (N-Caltech 101, N-Cars and EvTouch-Objects), and the following comparison is carried out in the evaluation process:
first, the results of the operation of the proposed method and the benchmark method of the present invention on two visual object recognition data sets (NCaltech 101 and N-Cars) are compared. Benchmark methods for comparison include HATS, matrix-LSTM, E2VID, event Frame), event Count Image, voxel Grid, and EST. All baseline methods use ResNet-34 as a classifier except E2VID (ResNet-18 is used). It can be seen from table 1 that the test accuracy of the present invention is much higher on both data sets than on all baseline methods.
Table 1N-Caltech 101 and N-Cars data set comparison of the results of the operation of the present invention with reference methods
Method N-Caltech101 (variance) N-Cars (variance)
HATS 69.7 90.9
EST 81.7 92.5
Matrix-LSTM 83.5(1.2) 92.2(0.7)
Event Frame 80.8(0.6) 93.4(0.5)
Event Count Image 81.1(0.4) 93.5(0.6)
Voxel Grid 83.9(0.3) 92.9(0.6)
MDOE (invention) 85.8(0.8) 94.2(0.4)
E2VD 86.6 91.0
MDOE (invention) 88.2(0.4) 94.6(0.3)
Secondly, comparing the results of the operation of the present method with the baseline method on the Evtouch-Objects data set, since Evtouch-Objects data is collected using irregularly placed tactile sensors, it is not possible to directly process these data using convolutional neural networks. To utilize convolutional neural networks, where haptic data is organized into an 11×11 Grid structure, the baseline method of comparison includes TactileSGNet, event Frame, event Count Image, voxel Grid, and EST. All framework-like methods and proposed MDOEs use res net-34 as classifier and use the same configuration (e.g., training time, learning rate). It can be seen from table 2 that the proposed method is significantly better than all reference methods.
TABLE 2 comparison of the results of the operation of the present invention with the benchmark method on the EvTouch-Objects dataset
Method Average accuracy (variance)
TactileSGNet 89.4(0.6)
EventFrame 62.9(1.1)
EventCountImage 62.9(1.6)
VoexlGrid 94.3(1.5)
EST 93.1(1.6)
MDOE (invention) 95.8(1.1)
Comparing the results of the method with the baseline method on the time delay, which is the period of time spent accumulating evidence in making decisions on object classes, is critical for applications that require fast reaction times. The data set evaluated was N-Caltech101, with the benchmark methods being Event Frame, event Count Image and Voxel Grid. It can be seen from fig. 3 that the present invention is superior to all baseline methods, and that an accuracy of about 82% can be achieved using only the first 30 milliseconds of the sample. This makes the invention well suited for applications with high real-time requirements, such as autonomous navigation.
Finally, comparing the operation results of the method with the operation results of the reference method when training data with different proportions are used, wherein the data set is N-Caltech101, and the reference method comprises Event frames, event Count Image and Voxel Grid. The present invention and all benchmark methods use the same configuration (e.g., resNet-34 network, random seed, and learning rate) for fairness. As can be seen from fig. 4, the classification accuracy of all methods increases with increasing ratio of training data used, and the proposed method is always more accurate than the reference. It can be seen that the proposed method achieves an accuracy of about 63% when only 10% of the training data is used, much higher than the Voxel Grid (about 59%), event Count Image (55%) and Event Frame (53%). This means that the proposed method can better handle small data sets and this feature is very advantageous in situations where it is difficult to collect large amounts of data.
Although the comparative implementation is based on object classification datasets, the proposed method may also be used for other event-based applications such as visual odometry, image reconstruction and optical flow estimation.
In this embodiment, an Adam optimizer is used to train the model by minimizing cross entropy loss, the initial learning rate is set to 0.0001, the value decays 0.5 times every 10 iterations in training, and the total number of iterations is set to 200. The batch size for all data sets was set to 4, model multiple runs were performed on each data set using different random number seeds for robust evaluation, and mean and standard deviation values were reported.
Finally, an ablation study was performed on the proposed method, first evaluating the effect of each component of the proposed MDOE, first comparing the object recognition accuracy using MDOE, MOE and DOE, respectively. ResNet-34 is used as a classifier and all methods use the same configuration (e.g., training time, learning rate). All methods use a 9 time box. As can be seen from Table 3, MDOE performs better than MOE and DOE.
TABLE 3 comparison of test accuracy for MOE, DOE, MDOE three methods
Representation method N-Caltech101 (variance) N-Cars (variance)
MOE 84.9(0.9) 93.5(0.5)
DOE 85.6(0.4) 92.9(0.5)
MDOE 85.8(0.8) 94.2(0.4)
Specifically, MOE is more accurate than DOE on N-Cars, but less accurate on N-Caltech101 than DOE. The proposed method of combining MOE and DOE is superior to the method of using MOE and DOE alone on both data sets. The effect of different time bins on the performance of the method of the invention was analyzed, with the number of time bins varying from 1 to 64. Table 4 shows that the number of time bins has an effect on the proposed method.
Table 4 comparison of test accuracy on different number of time bins
Time box N-Caltech101 (variance) N-Cars (variance)
1 83.4(0.4) 93.6(0.3)
9 85.8(0.8) 94.2(0.4)
16 86.2(0.7) 94.1(0.4)
32 86.6(0.8) 94.2(0.4)
64 86.4(0.6) 94.3(0.4)
When the time bin rises from 1 to 9, the accuracy achieved on N-Caltech101 increases from 83.4% to 85.8% and on N-Cars from 93.6% to 94.2%. The test accuracy continues to increase until the time bin of N-Caltech101 reaches 32. However, the situation is not the same for the N-Cars dataset, and the number of time bins after 9 has been reached has negligible effect on the proposed method. Next, the time bin was set to 9 to analyze the effect of other factors, and four methods proposed in the most advanced deep learning models, namely, res net-34, VGG-19, mobileNet-V2, and acceptance-V3, were evaluated. Table 5 shows the test accuracy on different network structures using the proposed method on the N-Caltech101 and N-Cars datasets. It can be seen that the accuracy of MobileNet-V2 and ResNet-34 is higher than VGG-19 and acceptance-V3 on both data sets.
TABLE 5 test accuracy on different models using MDOE
Model N-Caltech101 (variance) N-Cars (variance)
VGG-19 81.9(0.8) 91.9(0.6)
Inception-V3 84.9(0.6) 93.6(0.8)
MobileNet-V2 85.3(0.6) 94.8(0.8)
ResNet-34 85.8(0.8) 94.2(0.4)
Specifically, resNet-34 performs best on N-Caltech101, followed by MobileNet-V2 and acceptance-V3. In contrast, mobileNet V2 achieves the highest precision on N-Cars, followed by ResNet-34 and Incenpel-V3. VGG-19 achieves the lowest accuracy over both data sets.
Finally, it is noted that the above embodiments are only for illustrating the technical solution of the present invention and not for limiting the same, and although the present invention has been described in detail with reference to the preferred embodiments, it should be understood by those skilled in the art that the technical solution of the present invention may be modified without departing from the spirit and scope of the technical solution, and all such modifications are included in the scope of the claims of the present invention.

Claims (2)

1. A data representation method based on asynchronous event streams, characterized in that: the method comprises the following steps: s1, inputting an asynchronous event stream as event data; s2, acquiring event size MOE; s3, acquiring an event density DOE; s4, splicing the MOE and the DOE to obtain MDOE expression, wherein the obtained MDOE simultaneously comprises size information of event data and density information of the event;
in step S2, the sampling method is to set a fixed time interval when MOE is acquired, and sample at each fixed time interval to obtain a representation with dimensions of 2 xcxhxw, where H and W are the spatial height and width of the event image data, respectively, and C is the number of time bins;
in step S3, the event count over the whole event period is not directly considered in the process of obtaining the event density DOE, but rather the event count over each time window is used to generate a vector of size C for each pixel, and by doing this for each polarity, a DOE representation of size 2 xcxhxw is obtained;
in step S1, let epsilon be an asynchronous event input sequence expressed as:
Figure FDA0004136498570000011
wherein x is i Is the position, t i Is the timestamp of event generation, p i Is the polarity of the event, which takes two values: 1 and-1, respectively representing positive and negative events; i is the event number and the dynamics for impulse neurons are described as follows:
Figure FDA0004136498570000012
where u (t) represents the intimal potential of the neuron at time t,
Figure FDA0004136498570000013
is a weighted sum of the pre-neuron inputs, τ is a time constant, at a segmentIn time, the denser the events, the greater the membrane potential;
in steps S2 and S3, the MOE and DOE are acquired by sampling the asynchronous event sequence, respectively, assuming that E represents the MOE and V represents the DOE, which is described as follows:
Figure FDA0004136498570000014
Figure FDA0004136498570000015
k(a)=max(0,1-|a|) (5)
Figure FDA0004136498570000016
t n =t 1 +(c n +1)ΔT (8)
wherein ε is + And epsilon - A sequence of events having positive and negative polarity, respectively; k (a) is a bilinear kernel that mimics neuronal dynamics, 1 ei Is an indicator function that ensures that only events between specified time periods are counted; (x) l ,y m ,c n ) Is the space-time coordinates, x, on the voxel grid l ∈{0,1…,W-1},y m E {0,1, …, H-1}, c n E {0,1, …, N-1}; Δt is the size of the time bin; to reduce the effect of sensor noise, the resulting MOE and DOE are normalized by dividing them by their respective maximum values over the pixel and time bins;
in step S4, the MDOE is a splice of the MOE and DOE along the polarity, denoted by S, i.e
S(p,x l ,y m ,c n )=Cat(E(x l ,y m ,c n ),V(x l ,y m ,c n )) (9)
Where Cat denotes the splice operation, p denotes the polarity of the MOE and DOE, p e {0,1, 2, 3}, where p=0 and p=1 denote the on-off polarity of the MOE, and p=2 and p=3 correspond to the on-off polarity of the DOE.
2. A data representation system based on an asynchronous event stream, characterized by: the system employs the method of claim 1 for data representation of asynchronous event streams.
CN202210379153.6A 2022-04-12 2022-04-12 Data representation method and system based on asynchronous event stream Active CN114723009B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210379153.6A CN114723009B (en) 2022-04-12 2022-04-12 Data representation method and system based on asynchronous event stream

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210379153.6A CN114723009B (en) 2022-04-12 2022-04-12 Data representation method and system based on asynchronous event stream

Publications (2)

Publication Number Publication Date
CN114723009A CN114723009A (en) 2022-07-08
CN114723009B true CN114723009B (en) 2023-04-25

Family

ID=82243957

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210379153.6A Active CN114723009B (en) 2022-04-12 2022-04-12 Data representation method and system based on asynchronous event stream

Country Status (1)

Country Link
CN (1) CN114723009B (en)

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108182670A (en) * 2018-01-15 2018-06-19 清华大学 A kind of resolution enhancement methods and system of event image
EP3396550A1 (en) * 2017-04-28 2018-10-31 Ebury Technology Limited Asynchronous event-based instruction processing system and method
CN110991602A (en) * 2019-09-08 2020-04-10 天津大学 Event-driven pulse neuron simulation algorithm based on single exponential kernel
CN111695681A (en) * 2020-06-16 2020-09-22 清华大学 High-resolution dynamic visual observation method and device
CN112597980A (en) * 2021-03-04 2021-04-02 之江实验室 Brain-like gesture sequence recognition method for dynamic vision sensor
CN112712170A (en) * 2021-01-08 2021-04-27 西安交通大学 Neural morphology vision target classification system based on input weighted impulse neural network
CN113177640A (en) * 2021-05-31 2021-07-27 重庆大学 Discrete asynchronous event data enhancement method
CN113269699A (en) * 2021-04-22 2021-08-17 天津(滨海)人工智能军民融合创新中心 Optical flow estimation method and system based on fusion of asynchronous event flow and gray level image
CN114115152A (en) * 2021-11-25 2022-03-01 武汉智能装备工业技术研究院有限公司 Manufacturing edge real-time event insight method based on embedded type and deep learning

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3396550A1 (en) * 2017-04-28 2018-10-31 Ebury Technology Limited Asynchronous event-based instruction processing system and method
CN108182670A (en) * 2018-01-15 2018-06-19 清华大学 A kind of resolution enhancement methods and system of event image
CN110991602A (en) * 2019-09-08 2020-04-10 天津大学 Event-driven pulse neuron simulation algorithm based on single exponential kernel
CN111695681A (en) * 2020-06-16 2020-09-22 清华大学 High-resolution dynamic visual observation method and device
CN112712170A (en) * 2021-01-08 2021-04-27 西安交通大学 Neural morphology vision target classification system based on input weighted impulse neural network
CN112597980A (en) * 2021-03-04 2021-04-02 之江实验室 Brain-like gesture sequence recognition method for dynamic vision sensor
CN113269699A (en) * 2021-04-22 2021-08-17 天津(滨海)人工智能军民融合创新中心 Optical flow estimation method and system based on fusion of asynchronous event flow and gray level image
CN113177640A (en) * 2021-05-31 2021-07-27 重庆大学 Discrete asynchronous event data enhancement method
CN114115152A (en) * 2021-11-25 2022-03-01 武汉智能装备工业技术研究院有限公司 Manufacturing edge real-time event insight method based on embedded type and deep learning

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
刘彤 ; 倪维健 ; 孙宇健 ; 曾庆田 ; .基于深度迁移学习的业务流程实例剩余执行时间预测方法.数据分析与知识发现.2020,(第Z1期),全文. *
王晓浪 ; 邓蔚 ; 胡峰 ; 邓维斌 ; 张清华 ; .基于序列标注的事件联合抽取方法.重庆邮电大学学报(自然科学版).2020,(第05期),全文. *
闻佳 ; 王宏君 ; 邓佳 ; 刘鹏飞 ; .基于深度学习的异常事件检测.电子学报.2020,(第02期),全文. *
陈浩 ; 李永强 ; 冯远静 ; .基于多关系循环事件的动态知识图谱推理.模式识别与人工智能.2020,(第04期),全文. *
齐华青 ; .基于深度学习和稀疏组合的异常事件检测方法.电子测量技术.2019,(第20期),全文. *

Also Published As

Publication number Publication date
CN114723009A (en) 2022-07-08

Similar Documents

Publication Publication Date Title
Chung et al. An efficient hand gesture recognition system based on deep CNN
Hasan et al. Learning temporal regularity in video sequences
CN108596327B (en) Seismic velocity spectrum artificial intelligence picking method based on deep learning
CN112597980B (en) Brain-like gesture sequence recognition method for dynamic vision sensor
Dewan et al. Deeptemporalseg: Temporally consistent semantic segmentation of 3d lidar scans
CN110532959B (en) Real-time violent behavior detection system based on two-channel three-dimensional convolutional neural network
CN115601403A (en) Event camera optical flow estimation method and device based on self-attention mechanism
CN110610210A (en) Multi-target detection method
CN115146842B (en) Multi-element time sequence trend prediction method and system based on deep learning
CN111985333A (en) Behavior detection method based on graph structure information interaction enhancement and electronic device
Seidel et al. NAPC: A neural algorithm for automated passenger counting in public transport on a privacy-friendly dataset
Liu et al. Design of face detection and tracking system
CN114723009B (en) Data representation method and system based on asynchronous event stream
Xu et al. Tackling small data challenges in visual fire detection: a deep convolutional generative adversarial network approach
CN110633741B (en) Time sequence classification method based on improved impulse neural network
CN113076808A (en) Method for accurately acquiring bidirectional pedestrian flow through image algorithm
CN116843662A (en) Non-contact fault diagnosis method based on dynamic vision and brain-like calculation
Mahabal et al. Real-time classification of transient events in synoptic sky surveys
CN116229323A (en) Human body behavior recognition method based on improved depth residual error network
CN115964258A (en) Internet of things network card abnormal behavior grading monitoring method and system based on multi-time sequence analysis
Li et al. Research on YOLOv3 pedestrian detection algorithm based on channel attention mechanism
CN114035687A (en) Gesture recognition method and system based on virtual reality
Gu et al. MDOE: A spatiotemporal event representation considering the magnitude and density of events
Kong et al. One-Dimensional Convolutional Neural Networks Based on Exponential Linear Units for Bearing Fault Diagnosis
CN117079416B (en) Multi-person 5D radar falling detection method and system based on artificial intelligence algorithm

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant