EP0310636A1 - Distance measurement control of a multiple detector system. - Google Patents
Distance measurement control of a multiple detector system.Info
- Publication number
- EP0310636A1 EP0310636A1 EP88901684A EP88901684A EP0310636A1 EP 0310636 A1 EP0310636 A1 EP 0310636A1 EP 88901684 A EP88901684 A EP 88901684A EP 88901684 A EP88901684 A EP 88901684A EP 0310636 A1 EP0310636 A1 EP 0310636A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- value
- voiced
- frames
- calculating
- unvoiced
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000005259 measurement Methods 0.000 title claims abstract description 18
- 238000000926 separation method Methods 0.000 claims description 16
- 238000004364 calculation method Methods 0.000 claims description 15
- 238000000034 method Methods 0.000 claims description 13
- 239000013598 vector Substances 0.000 description 50
- 239000011159 matrix material Substances 0.000 description 9
- 238000012545 processing Methods 0.000 description 4
- 238000010586 diagram Methods 0.000 description 2
- 238000009826 distribution Methods 0.000 description 2
- 238000012549 training Methods 0.000 description 2
- 238000012935 Averaging Methods 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 230000015556 catabolic process Effects 0.000 description 1
- 238000006731 degradation reaction Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/93—Discriminating between voiced and unvoiced parts of speech signals
Definitions
- This invention relates to determining whether or not speech has a fundamental frequency present. This is also referred to as a voicing decision. More particularly, the invention is directed to selecting one of a plurality of voiced detectors which are concurrently processing speech samples for making the voicing decision with the selection being based on a distance measurement calculation.
- a voiced decision apparatus that selects between a plurality of voiced detectors by comparing separation or merit values generated by each of the voiced detectors.
- the separation values are also referred to as distance measurements.
- the apparatus comprises different types of voiced detectors such as discriminant and statistical detectors each generating a separation value.
- a comparator within the apparatus selects the voiced detector to make the determination whether the speech is voiced or unvoiced that is generating the largest separation value.
- the separation value may be a statistical, generalized distance value.
- All of the voiced detectors indicate whether a frame is voiced or unvoiced and each of the detectors first determines a discriminant variable for each, one of the present and previous frames. After determining the variable, each of the detectors determines mean values for both voiced and unvoiced ones of the previous and present frames. Each detector determines variance values for voiced and unvoiced ones of the previous and present frames. After calculating the means and the variances, each detector determines the separation value from the mean and variance values for the voiced frames and the mean and variance values for the unvoiced frames.
- the determination of the separation values is performed in each detector by combining variance values into a weighted sum.
- the mean value of each of the unvoiced frames is subtracted from the mean value of each of the voiced frames. This subtracted value is squared for each of the frames and the weighted sum of the variance values is divided into the resulting squared subtracted value.
- each detector before forming the weighted sum, each detector multiples the variance value for the voiced frames by the probability of a voiced frame occurring, and multiples the variance value for the unvoiced frames by the probability of an unvoiced frame occurring.
- the squared subtracted value is multiplied by the probabilities of an voiced frame occurring and unvoiced frame occurring.
- the method comprises the steps of calculating a first merit value defining the separation between voiced and unvoiced frames by the discriminant detector, calculating a second merit value defining separation between voiced and unvoiced frames by said statistical voiced detector, and selecting the detector that calculated best merit value to indicate whether a frame is voiced or unvoiced.
- FIG. 1 is a block diagram illustrating the present invention
- FIG. 2 illustrates, in block diagram form, statistical voice detector 103 of FIG. 1;
- FIGS. 3 and 4 illustrate, in greater detail, the functions performed by statistical voiced detector 103 of FIG. 2;
- FIG. 5 illustrates, in greater detail, functions performed by block 340 of FIG. 4.
- FIG. 1 illustrates an apparatus for performing the unvoiced/voiced decision operation by selecting between one of two voiced detectors. It would obvious to one skilled in the art to use more than two voiced detectors in FIG. 1.
- the selection between detectors 102 and 103 is based on a distance measurement that is generated by each detector and transmitted to distance comparator 104. Each generated distance measurement represents a merit value indicating the correctness of the generating detector's voicing decision.
- Distance comparator 104 compares the two distance measurement values and controls a multiplexer 105 such that the detector generating the greatest distance measurement value is selected to make the unvoiced/voiced decision. However, for other types of measurements, the lowest merit value would indicate the detector making the most accurate voicing decision.
- the distance measurement may be the Mahalanobis distance.
- detector 102 is a discriminant detector
- detector 103 is a statistical detector.
- the detectors could all be of the same type and that there could be more than two detectors present in the system.
- Classifier generator 101 is responsive to each frame of speech to generate classifiers which advantageously may be the log of the speech energy, the log of the LPC gain, the log area ratio of the first reflection coefficient, and the squared correlation coefficient of two speech segments one frame long which are offset by one pitch period.
- classifiers advantageously may be the log of the speech energy, the log of the LPC gain, the log area ratio of the first reflection coefficient, and the squared correlation coefficient of two speech segments one frame long which are offset by one pitch period.
- the calculation of these classifiers involves digitally sampling analog speech, forming frames of the digital samples, and processing those frames and is well known in the art.
- Generator 101 transmits the classifiers to detectors 102 and 103 via path 106.
- Detectors 102 and 103 are responsive to the classifiers received via path 106 to make unvoiced/voiced decisions and transmit these decisions via paths 107 and 110, respectively, to multiplexer 105.
- the detectors determine a distance measure between voiced and unvoiced frames and transmit these distances via paths 108 and 109 to comparator 104.
- these distances may be Mahalanobis distances or other generalized distances.
- Comparator 104 is responsive to the distances received via paths 108 and 109 to control multiplexer 105 so that the latter multiplexer selects the output of the detector that is generating the largest distance.
- FIG. 2 illustrates, in greater detail, statistical voiced detector 103.
- a set of classifiers also referred to as a vector of classifiers is received via path 106 from classifier generator 101.
- Silence detector 201 is responsive to these classifiers to determine whether or not speech is present in the present frame. If speech is present, detector 201 transmits a signal via path 210. If no speech (silence) is present in the frame, then only subtracter 207 and U/V determinator 205 are operational for that particular frame. Whether speech is present or not, the unvoiced/voiced decision is made for every frame by determinator 205.
- classifier averager 202 In response to the signal from detector 201, classifier averager 202 maintains an average of the individual classifiers received via path 106 by averaging in the classifiers for the present frame with the classifiers for previous frames. If speech (non-silence) is present in the frame, silence detector 201 signals statistical calculator 203, generator 206, and averager 202 via path 210. Statistical calculator 203 calculates statistical distributions for voiced and unvoiced frames. In particular, calculator 203 is responsive to the signal received via path 210 to calculate the overall probability that any frame is unvoiced and the probability that any frame is voiced.
- statistical calculator 203 calculates the statistical value that each classifier would have if the frame was unvoiced and the statistical value that each classifier would have if the frame was voiced. Further, calculator 203 calculates the covariance matrix of the classifiers. Advantageously, that statistical value may be the mean. The calculations performed by calculator 203 are not only based on the present frame but on previous frames as well. Statistical calculator 203 performs these calculations not only on the basis of the classifiers received for the present frame via path 106 and the average of the classifiers received path 211 but also on the basis of the weight for each classifiers and a threshold value defining whether, a frame is unvoiced or voiced received via path 213 from weights calculator 204.
- Weights calculator 204 is responsive to the probabilities, covariance matrix, and statistical values of the classifiers for the present frame as generated by calculator 203 and received via path 212 to recalculate the values used as weight vector a, for each of the classifiers and the threshold value b, for the present frame. Then, these new values of a and b are transmitted back to statistical calculator 203 via path 213. Also, weights calculator 204 transmits the weights and the statistical values for the classifiers in both the unvoiced and voiced regions via path 214, determinator 205, and path 208 to generator 206. The latter generator is responsive to this information to calculate the distance measure which is subsequently transmitted via path 109 to comparator 104 as illustrated in FIG. 1.
- U/V determinator 205 is responsive to the information transmitted via paths 214 and 215 to determine whether or not the frame is unvoiced or voiced and to transmit this decision via path 110 to multiplexer 105 of FIG. 1.
- Averager 202, statistical calculator 203, and weights calculator 204 implement an improved EM algorithm similar to that suggested in the article by N. E. Day entitled “Estimating the Components of a Mixture of Normal Distributions", Biometrika, Vol. 56, no. 3, pp. 463-474, 1969. Utilizing the concept of a decaying average, classifier averager 202 calculates the average for the classifiers for the present and previous frames by calculating following equations 1, 2, and 3:
- n n+1 if n ⁇ 2000 (1)
- z 1/n (2)
- x n is a vector representing the classifiers for the present frame, and n is the number of frames that have been processed up to 2000.
- z represents the decaying average coefficient
- X n represents the average of the classifiers over the present and past frames.
- Statistical calculator 203 is responsive to receipt of the z, x n and X n information to calculate the covariance matrix, T, by first calculating the matrix of sums of squares and products, Q n> as follows:
- T is calculated as follows:
- the means are subtracted from the classifiers as follows:
- calculator 203 determines the probability that the frame represented by the present vector x n is unvoiced by solving equation 7 shown below where, advantageously, the components of vector a are initialized as follows: component corresponding to log of the speech energy equals 0.3918606, component corresponding to log of the LPC gain equals -0.0520902, component corresponding to log area ratio of the first reflection coefficient equals 0.5637082, and component corresponding to squared correlation coefficient equals 1.361249; and b initially equals -8.36454:
- calculator 203 determines the probability that the classifiers represent a voiced frame by solving the following:
- calculator 203 determines the overall probability that any frame will be unvoiced by solving equation 9 for p n :
- calculator 203 determines two vectors, u and v, which give the mean values of each classifier for both unvoiced and voiced type frames.
- Vectors u and v are the statistical averages for unvoiced and voiced frames, respectively.
- Vector u statistical average unvoiced vector, contains the mean values of each classifier if a frame is unvoiced; and vector v, statistical average voiced vector, gives the mean value for each classifier if a frame is voiced.
- Vector u for the present frame is solved by calculating equation 10
- vector v is determined for the present frame by calculating equation 11 as follows:
- u n (1-z) u n _ ! + z x n P(u l x n) p n - zx n (10)
- v n (1-z) v n _ ! + z x n P(v lx n) /(l-p n) - zx n (11)
- Calculator 203 now communicates the u and v vectors, T matrix, and probability p to weights calculator 204 via path 212.
- Weights calculator 204 is responsive to this information to calculate new values for vector a and scalar b. These new values are then transmitted back to statistical calculator 203 via path 213. This allows detector 103 to adapt rapidly to changing environments. Advantageously, if the new values for vector a and scalar b are not transmitted back to statistical calculator 203, detector 103 will continue to adapt to changing environments since vectors u and v are being updated. As will be seen, determinator 205 uses vectors u and v as well as vector a and scalar b to make the voicing decision. If n is greater than advantageously 99, vector a and scalar b are calculated as follows. Vector a is determined by solving the following equation:
- Scalar b is determined by solving the following equation:
- weights calculator 204 transmits vectors a, u, and v to block 205 via path 214. If the frame contained silence only equation 6 is calculated. Determinator 205 is responsive to this transmitted information to decide whether the present frame is voiced or unvoiced. If the element of vector ( v n - u n ) corresponding to power is positive, then, a frame is declared voiced if the following equation is true: a'x r - a , (u n +v n) /2 > 0 ; (14)
- Equation 14 can also be rewritten as:
- Equation 15 can also be rewritten as:
- Equations 14 and 15 represent decision regions for making the voicing decision.
- the log term of the rewritten forms of equations 14 and 15 can be eliminated with some change of performance.
- the element corresponding to power is the log of the speech energy.
- Generator 206 is responsive to the information received via path 214 from calculator 204 to calculate the distance measure, A, as follows.
- the discriminant variable, d is calculated by equation 16 as follows:
- P d is initially set to .5.
- equations 21 through 24 are solved as follows:
- the probability, P d , that determinator 205 will declare a frame unvoiced is calculated by the following equation:
- Equation 25 uses Hotelling's two-sample T 2 statistic to calculate the distance measure.
- the distance measure can also be the Mahalanobis distance which is given in the following equation:
- Discriminant detector 102 makes the unvoiced/voiced decision by transmitting information to multiplexer 105 via path 107 indicating a voiced frame if a'x+b > 0. If this condition is not true, then detector 102 indicates an unvoiced frame.
- the values for vector a and scalar b used by detector 102 are advantageously identical to the initial values of a and b for statistical voiced detector 103.
- FIGS. 3 and 4 illustrate, in greater detail, the operations performed by statistical voiced detector 103 of FIG.2.
- Blocks 302 and 300 implement blocks 202 and 201 of FIG. 2, respectively.
- Blocks 304 through 318 implement statistical calculator 203.
- Blocks 320 and 322 implement weights calculator 204, and blocks 326 through 338 implement block 205 of F1G.2.
- Generator 206 of FIG. 2 is implemented by block 340.
- Subtractor 207 is implemented by block 308 or block 324.
- Block 302 calculates the vector which represents the average of the classifiers for the present frame and all previous frames.
- Block 300 determines whether speech or silence is present in the present frame; and if silence is present in the present frame, the mean for each classifier is subtracted from each classifier by block 324 before control is transferred to decision block 326. However, if speech is present in the present frame, then the statistical and weights calculations are performed by blocks 304 through 322.
- the average vector is found in block 302.
- Second, the sums of the squares and products matrix is calculated in block 304. The latter matrix along with the vector X representing the mean of the classifiers for the present and past frames is then utilized to calculate the covariance matrix, T, in block 306.
- Block 310 calculates the probability that the present frame is unvoiced by utilizing the present weight vector a, the present threshold value b, and the classifier vector for the present frame, x n . After calculating the probability that the present frame is unvoiced, the probability that the present frame is voiced is calculated by block 312. Then, the overall probability, p n , that any frame will be unvoiced is calculated by block 314.
- Blocks 316 and 318 calculate two vectors: u and v.
- the values contained in vector u represent the statistical average values that each classifier would have if the frame were unvoiced.
- vector v contains values representing the statistical average values that each classifier would have if the frame were voiced.
- the actual vectors of classifiers for the present and previous frames are clustered around either vector u or vector v.
- the vectors representing the classifiers for the previous and present frames are clustered around vector u if these frames are found to be unvoiced; otherwise, the previous classifier vectors are clustered around vector v.
- control is transferred to block 322; otherwise, control is transferred to block 326.
- block 322 Upon receiving control, block 322 then calculates a new weight vector a and a new threshold value b.
- the vector a and value b are used in the next sequential frame by the preceding blocks in FIG. 3.
- N is required to be greater than infinity, vector a and scalar b will never be changed, and detector 103 will adapt solely in response to vectors v and u as illustrated in blocks 326 through 338.
- Blocks 326 through 338 implement u/v determinator 205 of FIG. 2.
- Block 326 determines whether the power term of vector v of the present frame is greater than or equal to the power term of vector u. If this condition is true, then decision block 328 is executed. The latter decision block determines whether the test for voiced or unvoiced is met If the frame is found to be voiced in decision block 328, then the frame is so marked as voiced by block 330 otherwise the frame is marked as unvoiced by block 332. If the power term of vector v is less than the power term of vector u for the present frame, blocks 334 through 338 function are executed and function in a similar manner. Finally, block 340 calculates the distance measure.
- FIG. 5 illustrates, in greater detail the operations performed by block 340 of FIG. 4.
- Decision block 501 determines whether the frame has been indicated as unvoiced or voiced by examining the calculations 330, 332, 336, or 338. If the frame has been designated as voiced, path 507 is selected.
- Block 510 calculates probability P d , and block 502 recalculates the mean, mi, for the voiced frames and block 503 recalculates the variance, j . , for voiced frames. If the frame was determined to be unvoiced, decision block 501 selects path 508.
- Block 509 recalculates probability P d
- block 504 recalculates mean, mo, for unvoiced frames
- block 505 recalculates the variance ko for unvoiced frames.
- block 506 calculates the distance measure by performing the calculations indicated.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Multimedia (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Computational Linguistics (AREA)
- Acoustics & Sound (AREA)
- Signal Processing (AREA)
- Transmission Systems Not Characterized By The Medium Used For Transmission (AREA)
- Measurement Of Velocity Or Position Using Acoustic Or Ultrasonic Waves (AREA)
- Radar Systems Or Details Thereof (AREA)
- Measurement Of Mechanical Vibrations Or Ultrasonic Waves (AREA)
- Time-Division Multiplex Systems (AREA)
- Testing Or Calibration Of Command Recording Devices (AREA)
Abstract
Description
Claims
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
AT88901684T ATE80488T1 (en) | 1987-04-03 | 1988-01-11 | DISTANCE MEASUREMENT CONTROL OF A MULTI-DETECTOR SYSTEM. |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US3429787A | 1987-04-03 | 1987-04-03 | |
US34297 | 1987-04-03 |
Publications (2)
Publication Number | Publication Date |
---|---|
EP0310636A1 true EP0310636A1 (en) | 1989-04-12 |
EP0310636B1 EP0310636B1 (en) | 1992-09-09 |
Family
ID=21875527
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP88901684A Expired - Lifetime EP0310636B1 (en) | 1987-04-03 | 1988-01-11 | Distance measurement control of a multiple detector system |
Country Status (8)
Country | Link |
---|---|
EP (1) | EP0310636B1 (en) |
JP (1) | JPH0795238B2 (en) |
AT (1) | ATE80488T1 (en) |
CA (1) | CA1336212C (en) |
DE (1) | DE3874471T2 (en) |
HK (1) | HK108993A (en) |
SG (1) | SG59693G (en) |
WO (1) | WO1988007740A1 (en) |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
AU696092B2 (en) * | 1995-01-12 | 1998-09-03 | Digital Voice Systems, Inc. | Estimation of excitation parameters |
JP3670217B2 (en) | 2000-09-06 | 2005-07-13 | 国立大学法人名古屋大学 | Noise encoding device, noise decoding device, noise encoding method, and noise decoding method |
JP4517045B2 (en) * | 2005-04-01 | 2010-08-04 | 独立行政法人産業技術総合研究所 | Pitch estimation method and apparatus, and pitch estimation program |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS60114900A (en) * | 1983-11-25 | 1985-06-21 | 松下電器産業株式会社 | Voice/voiceless discrimination |
JPS60200300A (en) * | 1984-03-23 | 1985-10-09 | 松下電器産業株式会社 | Voice head/end detector |
JPS6148898A (en) * | 1984-08-16 | 1986-03-10 | 松下電器産業株式会社 | Voice/voiceless discriminator for voice |
-
1988
- 1988-01-11 JP JP63501560A patent/JPH0795238B2/en not_active Expired - Fee Related
- 1988-01-11 WO PCT/US1988/000123 patent/WO1988007740A1/en active IP Right Grant
- 1988-01-11 DE DE8888901684T patent/DE3874471T2/en not_active Expired - Fee Related
- 1988-01-11 AT AT88901684T patent/ATE80488T1/en active
- 1988-01-11 EP EP88901684A patent/EP0310636B1/en not_active Expired - Lifetime
- 1988-03-29 CA CA000562766A patent/CA1336212C/en not_active Expired - Fee Related
-
1993
- 1993-05-07 SG SG596/93A patent/SG59693G/en unknown
- 1993-10-14 HK HK1089/93A patent/HK108993A/en not_active IP Right Cessation
Non-Patent Citations (3)
Title |
---|
ICASSP 86 Proceedings, IEEE-IECEJ-ASJ International Conference on Acoustics, Speech, and Signal Processing, 7-11 April 1986, Tokyo, Japan, volume 1 of 4, IEEE, (New York, US), D.P. Prezas et al.: "Fast and accurate pitch detection using pattern recognition and adaptive time-domain analysis", pages 109-112 see pages 110, 111: "Final voicing and pitch" * |
IEEE Transactions on Acoustics, Speech, and Signal Processing, volume ASSP-24, no. 3, June 1976, (New York, US), B.S. Atal et al.: "A pattern recognition approach to voiced-unvoiched-silence classification with applications to speech recognition", pages 201-212 see page 201, right-hand column, lines 14-26 * |
See also references of WO8807740A1 * |
Also Published As
Publication number | Publication date |
---|---|
DE3874471D1 (en) | 1992-10-15 |
AU1242988A (en) | 1988-11-02 |
HK108993A (en) | 1993-10-22 |
DE3874471T2 (en) | 1993-02-25 |
SG59693G (en) | 1993-07-09 |
WO1988007740A1 (en) | 1988-10-06 |
ATE80488T1 (en) | 1992-09-15 |
JPH0795238B2 (en) | 1995-10-11 |
EP0310636B1 (en) | 1992-09-09 |
JPH01502853A (en) | 1989-09-28 |
CA1336212C (en) | 1995-07-04 |
AU602957B2 (en) | 1990-11-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP0335521B1 (en) | Voice activity detection | |
US6314396B1 (en) | Automatic gain control in a speech recognition system | |
US5276765A (en) | Voice activity detection | |
US5216702A (en) | Nonintrusive speech level and dynamic noise measurements | |
US20020165713A1 (en) | Detection of sound activity | |
US4937870A (en) | Speech recognition arrangement | |
US5046100A (en) | Adaptive multivariate estimating apparatus | |
US4972490A (en) | Distance measurement control of a multiple detector system | |
US5007093A (en) | Adaptive threshold voiced detector | |
FI117953B (en) | Audio signal processing device | |
WO2001031638A1 (en) | Handling variable delay in objective speech quality assessment | |
CA1337708C (en) | Adaptive multivariate estimating apparatus | |
EP0310636B1 (en) | Distance measurement control of a multiple detector system | |
EP0309561B1 (en) | An adaptive threshold voiced detector | |
JP2002258881A (en) | Device and program for detecting voice | |
GB2213623A (en) | Phoneme recognition | |
Milosavljevic et al. | Estimation of nonstationary AR model using the weighted recursive least square algorithm | |
US20040148168A1 (en) | Method and device for automatically differentiating and/or detecting acoustic signals | |
EP1269462B1 (en) | Voice activity detection apparatus and method | |
JP2000115260A (en) | Method and instrument for measuring q value | |
Yamazaki et al. | An objective method for evaluating the quality of speech with code errors using pattern matching techniques |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AT BE DE FR GB IT NL |
|
17P | Request for examination filed |
Effective date: 19890328 |
|
17Q | First examination report despatched |
Effective date: 19910402 |
|
GRAA | (expected) grant |
Free format text: ORIGINAL CODE: 0009210 |
|
AK | Designated contracting states |
Kind code of ref document: B1 Designated state(s): AT BE DE FR GB IT NL |
|
REF | Corresponds to: |
Ref document number: 80488 Country of ref document: AT Date of ref document: 19920915 Kind code of ref document: T |
|
REF | Corresponds to: |
Ref document number: 3874471 Country of ref document: DE Date of ref document: 19921015 |
|
ET | Fr: translation filed | ||
ITF | It: translation for a ep patent filed | ||
PLBE | No opposition filed within time limit |
Free format text: ORIGINAL CODE: 0009261 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT |
|
26N | No opposition filed | ||
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: BE Payment date: 20000106 Year of fee payment: 13 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: AT Payment date: 20000330 Year of fee payment: 13 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: AT Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20010111 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: BE Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20010131 |
|
BERE | Be: lapsed |
Owner name: AMERICAN TELEPHONE AND TELEGRAPH CY Effective date: 20010131 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: FR Payment date: 20011221 Year of fee payment: 15 |
|
REG | Reference to a national code |
Ref country code: GB Ref legal event code: IF02 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: NL Payment date: 20020107 Year of fee payment: 15 Ref country code: GB Payment date: 20020107 Year of fee payment: 15 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: DE Payment date: 20020328 Year of fee payment: 15 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: GB Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20030111 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: NL Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20030801 Ref country code: DE Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20030801 |
|
GBPC | Gb: european patent ceased through non-payment of renewal fee | ||
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: FR Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20030930 |
|
NLV4 | Nl: lapsed or anulled due to non-payment of the annual fee |
Effective date: 20030801 |
|
REG | Reference to a national code |
Ref country code: FR Ref legal event code: ST |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: IT Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20050111 |