US20210329903A1 - Systems and methods for classifying flying insects - Google Patents

Systems and methods for classifying flying insects Download PDF

Info

Publication number
US20210329903A1
US20210329903A1 US17/366,993 US202117366993A US2021329903A1 US 20210329903 A1 US20210329903 A1 US 20210329903A1 US 202117366993 A US202117366993 A US 202117366993A US 2021329903 A1 US2021329903 A1 US 2021329903A1
Authority
US
United States
Prior art keywords
insect
phototransistor
light transmission
flying
recorded
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US17/366,993
Inventor
Agenor Mafra-Neto
Eamonn Keogh
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Isca Technologies Inc
ISCA TECHNOLOGIES Inc
University of California
Original Assignee
Isca Technologies Inc
ISCA TECHNOLOGIES Inc
University of California
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Isca Technologies Inc, ISCA TECHNOLOGIES Inc, University of California filed Critical Isca Technologies Inc
Priority to US17/366,993 priority Critical patent/US20210329903A1/en
Assigned to ISCA TECHNOLOGIES, INC reassignment ISCA TECHNOLOGIES, INC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MAFRA-NETO, AGENOR
Assigned to THE REGENTS OF THE UNIVERSITY OF CALIFORNIA reassignment THE REGENTS OF THE UNIVERSITY OF CALIFORNIA ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: KEOGH, Eamonn
Publication of US20210329903A1 publication Critical patent/US20210329903A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A01AGRICULTURE; FORESTRY; ANIMAL HUSBANDRY; HUNTING; TRAPPING; FISHING
    • A01MCATCHING, TRAPPING OR SCARING OF ANIMALS; APPARATUS FOR THE DESTRUCTION OF NOXIOUS ANIMALS OR NOXIOUS PLANTS
    • A01M1/00Stationary means for catching or killing insects
    • A01M1/02Stationary means for catching or killing insects with devices or substances, e.g. food, pheronones attracting the insects
    • A01M1/026Stationary means for catching or killing insects with devices or substances, e.g. food, pheronones attracting the insects combined with devices for monitoring insect presence, e.g. termites
    • AHUMAN NECESSITIES
    • A01AGRICULTURE; FORESTRY; ANIMAL HUSBANDRY; HUNTING; TRAPPING; FISHING
    • A01MCATCHING, TRAPPING OR SCARING OF ANIMALS; APPARATUS FOR THE DESTRUCTION OF NOXIOUS ANIMALS OR NOXIOUS PLANTS
    • A01M1/00Stationary means for catching or killing insects
    • A01M1/02Stationary means for catching or killing insects with devices or substances, e.g. food, pheronones attracting the insects
    • AHUMAN NECESSITIES
    • A01AGRICULTURE; FORESTRY; ANIMAL HUSBANDRY; HUNTING; TRAPPING; FISHING
    • A01MCATCHING, TRAPPING OR SCARING OF ANIMALS; APPARATUS FOR THE DESTRUCTION OF NOXIOUS ANIMALS OR NOXIOUS PLANTS
    • A01M1/00Stationary means for catching or killing insects
    • A01M1/02Stationary means for catching or killing insects with devices or substances, e.g. food, pheronones attracting the insects
    • A01M1/04Attracting insects by using illumination or colours
    • AHUMAN NECESSITIES
    • A01AGRICULTURE; FORESTRY; ANIMAL HUSBANDRY; HUNTING; TRAPPING; FISHING
    • A01MCATCHING, TRAPPING OR SCARING OF ANIMALS; APPARATUS FOR THE DESTRUCTION OF NOXIOUS ANIMALS OR NOXIOUS PLANTS
    • A01M1/00Stationary means for catching or killing insects
    • A01M1/10Catching insects by using Traps
    • A01M1/106Catching insects by using Traps for flying insects
    • AHUMAN NECESSITIES
    • A01AGRICULTURE; FORESTRY; ANIMAL HUSBANDRY; HUNTING; TRAPPING; FISHING
    • A01MCATCHING, TRAPPING OR SCARING OF ANIMALS; APPARATUS FOR THE DESTRUCTION OF NOXIOUS ANIMALS OR NOXIOUS PLANTS
    • A01M1/00Stationary means for catching or killing insects
    • A01M1/20Poisoning, narcotising, or burning insects
    • AHUMAN NECESSITIES
    • A01AGRICULTURE; FORESTRY; ANIMAL HUSBANDRY; HUNTING; TRAPPING; FISHING
    • A01MCATCHING, TRAPPING OR SCARING OF ANIMALS; APPARATUS FOR THE DESTRUCTION OF NOXIOUS ANIMALS OR NOXIOUS PLANTS
    • A01M1/00Stationary means for catching or killing insects
    • A01M1/22Killing insects by electric means
    • A01M1/226Killing insects by electric means by using waves, fields or rays, e.g. sound waves, microwaves, electric waves, magnetic fields, light rays
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01VGEOPHYSICS; GRAVITATIONAL MEASUREMENTS; DETECTING MASSES OR OBJECTS; TAGS
    • G01V8/00Prospecting or detecting by optical means
    • G01V8/10Detecting, e.g. by using light barriers
    • G01V8/20Detecting, e.g. by using light barriers using multiple transmitters or receivers

Definitions

  • the present disclosure relates generally to systems for identifying flying insects and more particularly to noninvasive systems that optically record the wingbeat frequency of flying insects and convert the optical record to a sound file for accurate characterization and analysis of the recorded flying insect.
  • Flying insects have been bother civilization since the beginning of time, whether by destroying agricultural crops, spreading blood borne diseases, or simply interfering with outdoor activities. In order to mitigate these nuisances, it can be critically important to first classify the flying insects that are present in a given location, so that they can be properly managed, and/or for use in entomological research. Having an inexpensive, noninvasive system capable of accurately classifying the flying insects would allow for numerous advances in agricultural and medical applications, as well as in pure entomological research applications.
  • insects For example, nocturnal insects have been forced to fly by tapping and prodding them under bright halogen lights; insects have been recorded in confined spaces or under extreme temperatures. In some cases, insects were tethered with string to confine them within the range of the microphone. It is hard to imagine that such insect handling could result in data which would generalize to insects in natural conditions.
  • insects in flight have the “sound” of their wing beats extracted by an optical sensor. This is achieved by shining a light (source) at a phototransistor (target) and measuring the change in voltage (or electrical resistance or other electrical property). With no insects present there is no change in voltage. As the flying insect interrupts the path of light from the source to the target, its shadow causes a fluctuation in light intensity which the phototransistor converts into a fluctuation in voltage (or other electrical property). Then, an analog-to-digital converter is used to convert these fluctuations into a digital signal (in essence a sound file).
  • frequency spectrum also known as the energy spectral density
  • This frequency spectrum may be regarded as the insect's signature. This signature may be truncated at both ends to remove spurious data that does not reflect information produced by the insect.
  • the signature may undergo processing to normalize to a canonical invariant representation. This may be achieved by having the system record the corresponding humidity, temperature, air pressure, and/or other relevant environmental conditions. These environmental conditions are then factored into the analysis to normalize the signature to standard laboratory conditions.
  • This canonical representation is then input into a classifier (for example, a Bayesian Classifier, but other classifiers may also be suitable). Because classifiers are sensitive to the prior probability of seeing a particular insect, that information may also be input into the classifier.
  • the circadian rhythm of the insect may also be input into the classifier.
  • one aspect of the present disclosure is the ability to create a circadian rhythm for any possible time of dawn and dusk, given any two distinct circadian rhythms for a particular insect for a particular time of dusk and dawn. This system may then be used to compute the probability of seeing a particular insect at a given time of day, for any given day of year, at any given location on Earth.
  • the signal produced by an insect depends on the air density (which itself depends on altitude, humidity, temperature and air pressure).
  • the present system may correct for such differences caused by differences in air density from normal laboratory conditions to reach vastly increased correct identifications as opposed to the prior art.
  • a system for classifying flying insects includes a power source in electrical connection with an optical insect detector subsystem.
  • the optical insect detector subsystem features a first illuminator that emits a first light transmission, a first phototransistor positioned to receive the first light transmission, a second illuminator that emits a second light transmission, a second phototransistor positioned to receive the second light transmission, and a mechanical divider positioned between the first and second illuminators.
  • the mechanical divider is positioned to block the first light transmission from reaching the second phototransistor and to block the second light transmission from reaching the first phototransistor.
  • the first and second illuminators are positioned a defined distance apart from each other.
  • the system further includes an analog-to-digital converter in electrical communication with the first and second phototransistors.
  • the analog-to-digital converter is configured to produce a first digital signal from a change in voltage of the first phototransistor and to produce a second digital signal from a change in voltage of the second phototransistor.
  • the system also includes a recording device configured to record the first and second digital signals and a classification subsystem configured to receive the first and second digital signals from the recording device and to output a flying insect identification based upon the first and second digital signals.
  • the power source may be a battery and the first and second illuminators may be configured to emit at a constant brightness. Further, the first and second light transmissions may be created so as to not interfere with insect behavior. For example, the first and second light transmissions may be invisible to insects and/or may emit no measurable heat.
  • the first and second illuminators may be light emitting diodes. More particularly, in certain embodiments, the light emitting diodes may emit a light with a wavelength of approximately 940 nm and may be positioned approximately 1 cm apart from each other.
  • the system may further include an environmental sensor subsystem.
  • This environmental sensor subsystem may feature at least one sensor configured to measure at least one environmental condition data point surrounding the system.
  • the environmental subsystem may be in electronic communication with the recording device, such that the recording device records the environmental condition data points measured by the sensors and relays said recorded data points to the classification subsystem.
  • the sensors may measure environmental conditions such as, but not limited to, the humidity surrounding the system, the temperature surrounding the system, the air pressure surrounding the system, the ambient light surrounding the system, the current date, the current time, and/or the current location of the system.
  • the recording device may record the first and second digital signals as an audio file. More particularly, in certain embodiments, the audio file is a stereo audio file and the first digital signal is recorded as a left track and the second digital signal is recorded as a right track.
  • the recording device may further record an amplitude spectrum of the first and second digital signals.
  • the system may further include a circadian rhythm subsystem.
  • This circadian rhythm subsystem may be configured to calculate a circadian rhythm of an insect intended to be identified by the system based upon at least one environmental condition data point measured by the environmental sensor subsystem.
  • the system may further include an actuation subsystem electronically connected to the classification subsystem and configured to actuate a device upon receiving an identification signal from the classification subsystem. Examples of devices that may be actuated include, but are not limited to, a laser configured to kill an insect based upon receiving the identification signal, an electromagnetic capture device configured to capture an insect based upon receiving the identification signal and/or a solenoid valve configured to open and release a chemical based upon receiving the identification signal.
  • the capture device may comprise a cup attached to an electromagnet or an electromagnetically activated door.
  • Exemplary chemicals released by the solenoid valve include insect attractants, insect repellents, pesticides, and larvicides.
  • the apparatus features a power source in electrical connection with an optical insect detector unit.
  • the optical insect detector unit contains a first light emitting diode that emits a first light transmission having a wavelength of approximately 940 nm, a first phototransistor positioned to receive the first light transmission, a second light emitting diode that emits a second light transmission having a wavelength of approximately 940 nm, and a second phototransistor positioned to receive the second light transmission.
  • the first and second light emitting diodes are positioned approximately one centimeter apart from each other.
  • the optical insect detector unit further includes a mechanical divider positioned between the first and second light emitting diodes, such that the mechanical divider blocks the first light transmission from reaching the second phototransistor and blocks the second light transmission from reaching the first phototransistor.
  • the apparatus further includes an analog-to-digital converter in electrical communication with the first and second phototransistors.
  • the analog-to-digital converter is configured to produce a first digital signal from a change in voltage of the first phototransistor and to produce a second digital signal from a change in voltage of the second phototransistor.
  • the apparatus also features a recording device configured to record the first and second digital signals as stereo tracks of an audio file and an environmental sensor unit.
  • the environmental sensor unit has at least one sensor configured to measure at least one environmental condition data point surrounding the system. Environmental conditions to be measured include, but are not limited to, humidity, temperature, air pressure, ambient light, current date, current time, and current location.
  • the environmental sensor unit is in electronic communication with the recording device, such that the recording device records the environmental condition data points measured by the sensors.
  • the apparatus also includes a circadian rhythm unit configured to calculate a synthetic circadian rhythm of the insect intended to be identified by the apparatus.
  • the synthetic circadian rhythm is based upon at least one environmental condition data point measured by the environmental sensor unit, such as the location and/or current date.
  • the apparatus also has a classification unit configured to receive the first and second digital signals from the recording device and the environmental condition data points.
  • the classification unit outputs a flying insect identification based upon the first and second digital signals, at least one environmental condition data point, and the calculated synthetic circadian rhythm.
  • the apparatus may be a standalone device or may be contained within or attached to an insect trap.
  • Another embodiment envisioned by the present disclosure includes methods of classifying flying insects. These methods include the steps of recording a first set of data of the life cycle of a flying insect at set environmental conditions utilizing a first day and night cycle. This data includes at least the insect's wingbeat frequency, amplitude spectrum, and circadian rhythm and recording a second set of data of the life cycle of the flying insect at set environmental conditions utilizing a different, second day and night cycle. Similarly, the data recorded includes at least the insect's wingbeat frequency, amplitude spectrum, and circadian rhythm. Then, a data model of the flying insect is created based upon the data recorded in the earlier steps. This data model is then input into a classifier. After that, environmental conditions are recorded where the flying insect is to be classified. A synthetic circadian rhythm is then created based upon the data model and the recorded environmental conditions.
  • a first signal containing the wingbeat frequency and amplitude spectrum of the flying insect to be classified is recorded and a second signal containing the wingbeat frequency and amplitude spectrum of the flying insect to be classified is also recorded.
  • the second signal is captured at a known distance from the first signal. By doing so, one is able to calculate the approximate velocity and flight direction of the insect.
  • the first and second signals are then processed to normalize the signals based on the synthetic circadian rhythm and the current environmental conditions. These processed signals are then input into the classifier, which classifies the insect based upon the inputted signals and the data model.
  • the method may be utilized to classify the insect based on species, sex, or wing variation. Furthermore, the environmental conditions may be used to compensate for air density differences between those of the model and those of current conditions, and the resulting changes to the amplitude spectrum currently recorded from those recorded in the initial data recording steps. Moreover, the method may further include operating an actuator based upon the classification results.
  • FIG. 1 is a schematic view of the systems and apparatuses of the present disclosure.
  • FIG. 2 is a flowchart of the methods described in the present disclosure.
  • One aspect of the present disclosure is a system that can non-evasively capture signals from flying insects, and use this information to classify the insects by species, physiological state, and/or sex (if the insect is sexually dimorphic).
  • This system can be a standalone device, or can be integrated into the entrance of, or within, a standard insect trap.
  • physiological state include both natural physiological states and those induced by researchers.
  • An example of a natural physiological state that can be differentiated by the presently disclosed systems is whether a female mosquito has or has not recently taken a blood meal.
  • an example of an induced physiological state is when a researcher has deliberately removed part of the wing(s) of an insect.
  • the optical insect detector subsystem 12 includes a plurality of light sources (or illuminators) 14 a,b and an equal number of phototransistors 16 a,b.
  • the optical insect detector subsystem 12 is electrically connected to, and powered by a, a power source 18 .
  • the power source 18 may be, for example, a battery.
  • the illuminators 14 a,b should produce a constant brightness.
  • a suitable light source is, for example, an LED that emits light at a wavelength of at, or around, 940 nm (infrared). This wavelength is not visible to insects (or humans) and does not produce any detectible heat.
  • a first illuminator 14 a and a first phototransistor 16 a face each other from some distance, such that some of the light 20 a that is emitted from the first illuminator 14 a falls upon the first phototransistor 16 a.
  • the optical insect detector subsystem 12 utilizes two illuminators 14 a,b and two corresponding phototransistors 16 a,b, which are placed in a parallel, side-by-side arrangement, some small distance apart, such that the light 20 a from the first illuminator 14 a only falls upon the first phototransistor 16 a, and light 20 b from a second illuminator 14 b falls only upon a second phototransistor 16 b.
  • the illuminators 14 a,b are placed approximately one centimeter apart. In the absence of insects, or anything else, that blocks part of the light 20 a,b, the amount of light 20 a,b that falls on the phototransistors 16 a,b is constant. Furthermore, by examining the timing difference between when an insect crosses the first light beam 20 a and the second light beam 20 b, the system 10 can infer the speed and direction of travel of the insect. The speed and direction of travel are two inputs that may be fed into the system's classification algorithms to identify the flying insect.
  • the phototransistors 16 a,b are connected to an analog to digital converter 22 , and in a constant light field, no signal 23 (not shown) is produced. However, when an object blocks all or part of the light 20 a,b from the illuminators 14 a,b, the phototransistors 16 a,b will create a changing voltage that the analog to digital converter 22 will convert to a signal 23 .
  • the signal 23 produced by the analog to digital converter 22 is then captured by a recording device 24 . While this signal 23 is recorded optically as described above, it is effectively a sound file, and can be saved in any standard computer sound file format (e.g., as a WAV file), and if played on a standard audio device sounds like a conventionally recorded sound file recorded with a microphone.
  • this embodiment utilizes two illuminators 14 a,b and two phototransistors 16 a,b, two separate signals 23 a,b will be produced by the analog to digital converter 22 , which can be recorded as separate stereo tracks by the recording device 24 .
  • the system 10 continuously records and monitors the two independent signals 23 a,b. In the absence of insects there should be no signal 23 produced. In practice, however, electronic devices always produce a tiny amount of signal due to noise in the circuit (for example, wires within the electronics may act as a weak antenna, picking up a tiny amount of radio signals or other electromagnetic interference). Accordingly, the system 10 will typically utilize a threshold to determine if the device is currently seeing a signal 23 caused by the presence of an insect or not. This threshold can be computed in several ways. For example, the device can be sealed so that there are no insects present, at which point the mean and standard deviation of the noise amplitude is measured. In one embodiment, the signal threshold is then set to be the measured mean plus four standard deviations.
  • the analog to digital converter 22 will produce signals 23 a,b whose amplitude greatly exceed the threshold.
  • the recording device 24 then records a sound snippet of the interception event.
  • the sound snippet is recorded from one second before the interception, to one second after the interception.
  • the system 10 is able to record a sound snippet of this event one second before it happens, by maintaining the signals 23 a,b in a circular buffer. That is, the recording device 24 is constantly recording the signals 23 a,b in a constant loop, and when an interception event occurs, the recording is tagged to begin at one second prior in the recording loop.
  • the system 10 further utilizes sensors 26 to measure environmental conditions surrounding the system 10 .
  • Examples of information that may be measured by the sensors 26 include, but are not limited to, the humidity, temperature, air pressure, ambient light, date, time, and location on Earth.
  • the sensors 26 are in electrical communication with the recording device 24 , such that information measured by the sensors 26 is sent to be recorded by the recording device 24 .
  • the recording device 24 is therefore able to create a holistic record of the insect interception event that may include, among other things, a stereo sound file containing the signals 23 a,b, assigned to specific stereo tracks; humidity; temperature; air pressure; ambient light; date; time; location on Earth, and/or the amplitude spectrum derived from the stereo sound file.
  • the amplitude spectrum is a vector of real numbers, and provides an additional important piece of information useful in differentiating one species from another that have similar wingbeat frequencies. By additionally recording and analyzing the amplitude spectrum, error rates in correctly identifying the species are dramatically reduced. This is similar to the scenario wherein an individual can differentiate middle C played on a piano and middle C played on a violin.
  • This information may be recorded in various sound file formats, for example WAV or MP3 formats.
  • the two signals 23 a,b may be recorded as stereo sound tracks and the remaining data may be embedded within the sound file as metadata.
  • This offline learning step 28 can be achieved by placing a system 10 into an insectary (not shown) with juvenile insects of a single species, and single sex and species for sexually dimorphic insects (i.e. pupa for mosquitoes, maggots for flies, bees and wasps, etc.) and any food, water or other items needed to sustain healthy adult insect.
  • the system 10 is operated continuously to record the emerging insects, 24 hours a day, for their entire adult lives.
  • This recording is conducted at a fixed humidity, temperature, and air pressure, with a fixed artificial day and night cycle (e.g., 16 hours of darkness, followed by 8 hours of light, with a five minute linear ramp-up/ramp-down of the brightness to simulate dawn and dusk).
  • a fixed artificial day and night cycle e.g., 16 hours of darkness, followed by 8 hours of light, with a five minute linear ramp-up/ramp-down of the brightness to simulate dawn and dusk.
  • a similar recording is conducted utilizing a different fixed artificial day and night cycle (e.g., 12 hours of darkness followed by 12 hours of light), with otherwise the same environmental conditions as the first recording.
  • Collecting data in the manner allows the creation of excellent insect models regarding the circadian rhythm 30 a of the insect, for those recorded environment conditions.
  • the real world does not necessarily operate under such ideal conditions.
  • the methods of the present disclosure allow for the generalization from the recorded conditions in the offline learning step 28 to any reasonable single environment condition 30 b, as shown in the flowchart of FIG. 2 .
  • the probability that a given insect will be flying at a given time of day can be encoded in a flight activity circadian rhythm.
  • Certain embodiments of systems and methods described herein may include a circadian rhythm subsystem 32 that can create an insect specific circadian rhythm for any day of the year, for any location on Earth, after seeing just two distinct example circadian rhythms for that insect in the offline learning step 28 . Because this prior probability can vary by orders of magnitude over the course of the day, it greatly affects the ability to correctly identify insects.
  • the daily flight rhythms of Culex tarsalis and Aedes aegypti are very different, and offer a useful feature for the Bayesian classifier used in one embodiment of the system 10 .
  • an insect detected at 3 am is about four times more likely to be a Cx. tarsalis, whereas an insect spotted at midday is almost certainly an Ae. aegypti.
  • the circadian rhythm subsystem 32 given two distinct circadian rhythms observed during the offline learning step 28 , can produce a high quality approximation to the circadian rhythm that would be observed for any timing of dawn and dusk via the circadian rhythm modeling step 30 a.
  • the system 10 creates a synthetic circadian rhythm in the circadian rhythm modeling step 30 a using a type of inverted Dynamic Time Warping (DTW) based on the information provided from the offline learning step 28 .
  • DTW inverted Dynamic Time Warping
  • the DTW algorithm takes as input two time series, C and Q, and produces a warping path in an alignment matrix as a side effect of calculating the distance DTW(Q,C).
  • the system 10 uses a modified DTW algorithm to align the two given two distinct circadian rhythms to the desired circadian rhythm (the latter specified by desired time of dawn and dusk).
  • the amount of x-axis “warping” needed to convert the observed circadian rhythms into the desired circadian rhythm is used as an input into a regression algorithm, which is then used to set the y-axis value of the circadian rhythm.
  • the circadian rhythm subsystem 32 takes the synthetic circadian rhythm produced in step 30 a and the captured signals 23 a,b and feeds this information to the classifier 34 .
  • This data is then utilized in an environmental conditions modeling step 30 b to create an environmental conditions subsystem 36 .
  • the captured signals 23 a,b and information provided from the environmental sensors 26 can be input into the environmental conditions subsystem 36 to render data that has been compensated based on the environmental conditions to match those of ideal test conditions.
  • the captured signals 23 a,b as modified by the environmental conditions subsystem 36 are then fed to the classifier 34 .
  • the information gathered from the offline learning step 28 including amplitude spectrum, wingbeat frequency, daily flight rhythms, average flight speed and/or direction, etc. is used to build the classifier 34 . While any classifier may be used, in certain embodiments a Bayesian Classifier is utilized.
  • the system 10 is placed in the desired location and switched on. After the insect signals 23 a,b are captured, the system 10 normalizes the signals 23 a,b to achieve invariance to ambient altitude, humidity, temperature and air pressure with the environmental conditions subsystem 36 , and to the current time of dawn/dusk with the circadian rhythm subsystem 32 . The output from these subsystems 32 , 36 is then input into the classifier 34 , which produces the predicted class label of the sex and/or species and/or physiological state.
  • the system 10 may optionally use the ability to classify a single insect, or the cumulative number/ratios of classified insects observed thus far, to control an actuator 38 .
  • an actuator 38 can be used to selectively kill insects based on their sex. This is a potentially useful ability to support sterile insect technique (SIT) for mosquitoes. In such cases you want to release only males from a trap or insectary. However, other embodiments could allow for the actuator 38 to kill only a particular pest insect while allowing a beneficial insect to escape the trap unharmed.
  • SIT sterile insect technique
  • An exemplary kill method would utilize a powerful laser aimed at the exit of a trap or insectary, wherein the actuator 38 activates the laser to kill the desired insect.
  • the insect Upon attempting to exit the trap, the insect is classified via the system 10 , and if it is determined to be a target insect, the actuator 38 activates the laser. When a non-target insect is identified, the system 10 does not activate the actuator 38 .
  • the actuator 38 may be utilized to capture the target insect.
  • the system 10 is attached to an apparatus that can capture living insects, such as a simple plastic cup suspend over a flat surface.
  • the cup may be held up by an electromagnet connected to the actuator 38 .
  • a signal is sent to the actuator 38 to turn off the electromagnet, causing the cup to drop on top of the target insect, thereby capturing it.
  • the actuator 38 could close a door in a trap, thereby retaining the target insect inside the trap.
  • the system 10 is used to control at least one valve solenoid.
  • the solenoid may be opened by the actuator 38 to release chemicals in response to the identification of a single insect, or the cumulative number/ratios of identified insects observed thus far. Examples of chemicals that may be released include, but are not limited to, attractants, repellents, pesticides or larvicides.
  • Agricultural: Insect pests destroy billions of dollars worth of food each year. Any attempt to mitigate their harmful effects requires knowing the density of the species present (and sometimes the sex of the insects). The present systems can provide this species/sex count in real time.
  • Vector Control Districts Most vector control districts spend considerable effort to quantify the density of various mosquitoes in their district. Using conventional methods, this information can be inaccurate, out-of-date (by days or even weeks) and expensive to obtain. The present systems can provide real-time accurate information at a very low amortized cost.

Abstract

Systems, apparatuses, and methods of classifying flying insects. The methods utilize recording the wingbeat frequency and amplitude spectrum of flying insects and comparing them to known or created insect models to properly classify the flying insect. The error rate of the classification is reduced by utilizing multiple inputs to the classification system, which may include a precise circadian rhythm for the time of year, current environmental conditions, and flight velocity or direction.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation of U.S. patent application Ser. No. 16/817,468, filed on Mar. 12, 2020, which was a continuation of U.S. patent application Ser. No. 16/221,199, filed on Dec. 14, 2018, which was a continuation of U.S. patent application Ser. No. 15/253,286, now issued as U.S. Pat. No. 10,178,856 on Jan. 15, 2019, titled SYSTEMS AND METHODS FOR CLASSIFYING FLYING INSECTS, which claimed the benefit of U.S. Provisional Application No. 62/212,744, filed on Sep. 1, 2015, titled A SYSTEM TO CLASSIFY THE SEX, SPECIES, AND PHYSIOLOGICAL STATE OF FLYING INSECTS, the teachings of which are all expressly incorporated by reference.
  • STATEMENT RE: FEDERALLY SPONSORED RESEARCH/DEVELOPMENT
  • Not Applicable
  • BACKGROUND Field of the Invention
  • The present disclosure relates generally to systems for identifying flying insects and more particularly to noninvasive systems that optically record the wingbeat frequency of flying insects and convert the optical record to a sound file for accurate characterization and analysis of the recorded flying insect.
  • Description of the Prior Art
  • Flying insects have been bother mankind since the beginning of time, whether by destroying agricultural crops, spreading blood borne diseases, or simply interfering with outdoor activities. In order to mitigate these nuisances, it can be critically important to first classify the flying insects that are present in a given location, so that they can be properly managed, and/or for use in entomological research. Having an inexpensive, noninvasive system capable of accurately classifying the flying insects would allow for numerous advances in agricultural and medical applications, as well as in pure entomological research applications.
  • The idea of automatically classifying insects using the incidental sound of their flight (as opposed to deliberate insect sounds produced by stridulation) dates back to the very dawn of computers and commercially available audio recording equipment. In 1945, three researchers at the Cornell University Medical College, Kahn, Celestin and Offenhauser, used equipment donated by Oliver E. Buckley (then President of the Bell Telephone Laboratories) to record and analyze mosquito sounds. These authors later wrote, “It is the authors' considered opinion that the intensive application of such apparatus will make possible the precise, rapid, and simple observation of natural phenomena related to the sounds of disease-carrying mosquitoes and should lead to the more effective control of such mosquitoes and of the diseases that they transmit.”
  • Since then, there have been sporadic efforts at flying insect classification from audio features, especially in the last decade; however, little real progress seems to have been made. The lack of prior progress may be due in part to the fact that most researchers have used acoustic microphones. Sound attenuates according to an inverse squared law. For example, if an insect flies just three times further away from the microphone, the sound intensity (informally, the loudness) drops to one ninth. Any attempt to mitigate this by using a more sensitive microphone invariably results in extreme sensitivity to wind noise and to ambient noise in the environment. Moreover, the difficulty of collecting data with such devices seems to have led some researchers to obtain data in unnatural conditions. For example, nocturnal insects have been forced to fly by tapping and prodding them under bright halogen lights; insects have been recorded in confined spaces or under extreme temperatures. In some cases, insects were tethered with string to confine them within the range of the microphone. It is hard to imagine that such insect handling could result in data which would generalize to insects in natural conditions.
  • Furthermore, the vast majority of attempts to classify insects by their flight sounds have explicitly or implicitly used just the wingbeat frequency. However, such an approach is limited to applications in which the insects to be discriminated have very different frequencies (such as a butterfly and a mosquito), and is not well suited for real work applications where there may be multiple insects present with similar wingbeat frequencies.
  • As such, there is a need for systems and methods for efficiently and economically classifying flying insects by species, sex, physiological state, and other characteristics. Moreover, these systems should be able to classify the flying insects in real-world situations with minimal error rates. Furthermore, there is a need for systems that are capable of identifying flying insects regardless of wherever in the world they may be used, at any point in time, and under any environmental condition.
  • BRIEF SUMMARY
  • The systems and methods described herein are capable of obtaining data from flying insects, and classifying them down to the level of the sex, species and/or physiological state. In order to achieve this, insects in flight have the “sound” of their wing beats extracted by an optical sensor. This is achieved by shining a light (source) at a phototransistor (target) and measuring the change in voltage (or electrical resistance or other electrical property). With no insects present there is no change in voltage. As the flying insect interrupts the path of light from the source to the target, its shadow causes a fluctuation in light intensity which the phototransistor converts into a fluctuation in voltage (or other electrical property). Then, an analog-to-digital converter is used to convert these fluctuations into a digital signal (in essence a sound file). Various sound features are extracted from the sound file, including the frequency spectrum (also known as the energy spectral density). This frequency spectrum may be regarded as the insect's signature. This signature may be truncated at both ends to remove spurious data that does not reflect information produced by the insect.
  • Because the signature is affected by the humidity/temperature and air pressure in which it was recorded, the signature may undergo processing to normalize to a canonical invariant representation. This may be achieved by having the system record the corresponding humidity, temperature, air pressure, and/or other relevant environmental conditions. These environmental conditions are then factored into the analysis to normalize the signature to standard laboratory conditions. This canonical representation is then input into a classifier (for example, a Bayesian Classifier, but other classifiers may also be suitable). Because classifiers are sensitive to the prior probability of seeing a particular insect, that information may also be input into the classifier.
  • As the prior probability of seeing an insect strongly depends on the time relative to dusk and dawn (i.e. the circadian rhythm), the circadian rhythm of the insect may also be input into the classifier. In particular, one aspect of the present disclosure is the ability to create a circadian rhythm for any possible time of dawn and dusk, given any two distinct circadian rhythms for a particular insect for a particular time of dusk and dawn. This system may then be used to compute the probability of seeing a particular insect at a given time of day, for any given day of year, at any given location on Earth.
  • Likewise, the signal produced by an insect depends on the air density (which itself depends on altitude, humidity, temperature and air pressure). By measuring these characteristics at the time of the recording, the present system may correct for such differences caused by differences in air density from normal laboratory conditions to reach vastly increased correct identifications as opposed to the prior art.
  • In accordance with one embodiment of the present disclosure, there is contemplated a system for classifying flying insects. The system includes a power source in electrical connection with an optical insect detector subsystem. The optical insect detector subsystem features a first illuminator that emits a first light transmission, a first phototransistor positioned to receive the first light transmission, a second illuminator that emits a second light transmission, a second phototransistor positioned to receive the second light transmission, and a mechanical divider positioned between the first and second illuminators. The mechanical divider is positioned to block the first light transmission from reaching the second phototransistor and to block the second light transmission from reaching the first phototransistor. Furthermore, the first and second illuminators are positioned a defined distance apart from each other.
  • The system further includes an analog-to-digital converter in electrical communication with the first and second phototransistors. The analog-to-digital converter is configured to produce a first digital signal from a change in voltage of the first phototransistor and to produce a second digital signal from a change in voltage of the second phototransistor. The system also includes a recording device configured to record the first and second digital signals and a classification subsystem configured to receive the first and second digital signals from the recording device and to output a flying insect identification based upon the first and second digital signals.
  • The power source may be a battery and the first and second illuminators may be configured to emit at a constant brightness. Further, the first and second light transmissions may be created so as to not interfere with insect behavior. For example, the first and second light transmissions may be invisible to insects and/or may emit no measurable heat. In certain embodiments, the first and second illuminators may be light emitting diodes. More particularly, in certain embodiments, the light emitting diodes may emit a light with a wavelength of approximately 940 nm and may be positioned approximately 1 cm apart from each other.
  • The system may further include an environmental sensor subsystem. This environmental sensor subsystem may feature at least one sensor configured to measure at least one environmental condition data point surrounding the system. Furthermore, the environmental subsystem may be in electronic communication with the recording device, such that the recording device records the environmental condition data points measured by the sensors and relays said recorded data points to the classification subsystem. The sensors may measure environmental conditions such as, but not limited to, the humidity surrounding the system, the temperature surrounding the system, the air pressure surrounding the system, the ambient light surrounding the system, the current date, the current time, and/or the current location of the system.
  • The recording device may record the first and second digital signals as an audio file. More particularly, in certain embodiments, the audio file is a stereo audio file and the first digital signal is recorded as a left track and the second digital signal is recorded as a right track. The recording device may further record an amplitude spectrum of the first and second digital signals.
  • The system may further include a circadian rhythm subsystem. This circadian rhythm subsystem may be configured to calculate a circadian rhythm of an insect intended to be identified by the system based upon at least one environmental condition data point measured by the environmental sensor subsystem. Additionally, or alternatively, the system may further include an actuation subsystem electronically connected to the classification subsystem and configured to actuate a device upon receiving an identification signal from the classification subsystem. Examples of devices that may be actuated include, but are not limited to, a laser configured to kill an insect based upon receiving the identification signal, an electromagnetic capture device configured to capture an insect based upon receiving the identification signal and/or a solenoid valve configured to open and release a chemical based upon receiving the identification signal. The capture device may comprise a cup attached to an electromagnet or an electromagnetically activated door. Exemplary chemicals released by the solenoid valve include insect attractants, insect repellents, pesticides, and larvicides.
  • Another embodiment contemplated by the present disclosure is an apparatus for classifying flying insects. The apparatus features a power source in electrical connection with an optical insect detector unit. The optical insect detector unit contains a first light emitting diode that emits a first light transmission having a wavelength of approximately 940 nm, a first phototransistor positioned to receive the first light transmission, a second light emitting diode that emits a second light transmission having a wavelength of approximately 940 nm, and a second phototransistor positioned to receive the second light transmission. The first and second light emitting diodes are positioned approximately one centimeter apart from each other. The optical insect detector unit further includes a mechanical divider positioned between the first and second light emitting diodes, such that the mechanical divider blocks the first light transmission from reaching the second phototransistor and blocks the second light transmission from reaching the first phototransistor.
  • The apparatus further includes an analog-to-digital converter in electrical communication with the first and second phototransistors. The analog-to-digital converter is configured to produce a first digital signal from a change in voltage of the first phototransistor and to produce a second digital signal from a change in voltage of the second phototransistor. The apparatus also features a recording device configured to record the first and second digital signals as stereo tracks of an audio file and an environmental sensor unit. The environmental sensor unit has at least one sensor configured to measure at least one environmental condition data point surrounding the system. Environmental conditions to be measured include, but are not limited to, humidity, temperature, air pressure, ambient light, current date, current time, and current location. The environmental sensor unit is in electronic communication with the recording device, such that the recording device records the environmental condition data points measured by the sensors.
  • The apparatus also includes a circadian rhythm unit configured to calculate a synthetic circadian rhythm of the insect intended to be identified by the apparatus. The synthetic circadian rhythm is based upon at least one environmental condition data point measured by the environmental sensor unit, such as the location and/or current date.
  • The apparatus also has a classification unit configured to receive the first and second digital signals from the recording device and the environmental condition data points. The classification unit outputs a flying insect identification based upon the first and second digital signals, at least one environmental condition data point, and the calculated synthetic circadian rhythm. The apparatus may be a standalone device or may be contained within or attached to an insect trap.
  • Another embodiment envisioned by the present disclosure includes methods of classifying flying insects. These methods include the steps of recording a first set of data of the life cycle of a flying insect at set environmental conditions utilizing a first day and night cycle. This data includes at least the insect's wingbeat frequency, amplitude spectrum, and circadian rhythm and recording a second set of data of the life cycle of the flying insect at set environmental conditions utilizing a different, second day and night cycle. Similarly, the data recorded includes at least the insect's wingbeat frequency, amplitude spectrum, and circadian rhythm. Then, a data model of the flying insect is created based upon the data recorded in the earlier steps. This data model is then input into a classifier. After that, environmental conditions are recorded where the flying insect is to be classified. A synthetic circadian rhythm is then created based upon the data model and the recorded environmental conditions.
  • A first signal containing the wingbeat frequency and amplitude spectrum of the flying insect to be classified is recorded and a second signal containing the wingbeat frequency and amplitude spectrum of the flying insect to be classified is also recorded. The second signal is captured at a known distance from the first signal. By doing so, one is able to calculate the approximate velocity and flight direction of the insect. The first and second signals are then processed to normalize the signals based on the synthetic circadian rhythm and the current environmental conditions. These processed signals are then input into the classifier, which classifies the insect based upon the inputted signals and the data model.
  • The method may be utilized to classify the insect based on species, sex, or wing variation. Furthermore, the environmental conditions may be used to compensate for air density differences between those of the model and those of current conditions, and the resulting changes to the amplitude spectrum currently recorded from those recorded in the initial data recording steps. Moreover, the method may further include operating an actuator based upon the classification results.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • These and other features and advantages of the various embodiments disclosed herein will be better understood with respect to the following description and drawings, in which like numbers refer to like parts throughout, and in which:
  • FIG. 1 is a schematic view of the systems and apparatuses of the present disclosure; and
  • FIG. 2 is a flowchart of the methods described in the present disclosure.
  • DETAILED DESCRIPTION
  • The detailed description set forth below is intended as a description of the presently preferred embodiment of the invention, and is not intended to represent the only form in which the present invention may be constructed or utilized. The description sets forth the functions and sequences of steps for constructing and operating the invention. It is to be understood, however, that the same or equivalent functions and sequences may be accomplished by different embodiments and that they are also intended to be encompassed within the scope of the invention.
  • One aspect of the present disclosure is a system that can non-evasively capture signals from flying insects, and use this information to classify the insects by species, physiological state, and/or sex (if the insect is sexually dimorphic). This system can be a standalone device, or can be integrated into the entrance of, or within, a standard insect trap. Examples of physiological state include both natural physiological states and those induced by researchers. An example of a natural physiological state that can be differentiated by the presently disclosed systems is whether a female mosquito has or has not recently taken a blood meal. Whereas an example of an induced physiological state, is when a researcher has deliberately removed part of the wing(s) of an insect. It is known in the art that a large area of insect wings can be removed with minimal effect on load-lifting ability. Furthermore, it is known that combinatorial wing damage patterns can be utilized to mark an insect. For example, the anal veins of a left wing on an insect may be removed, while the radius of a right wing of the insect may also be removed. Since the systems described herein are capable of differentiating the slight changes in flight behavior caused by such intentional marking of the insect's wings, they may be very useful in mark-and-recapture applications. Mark-and-recapture is frequently used to estimate, among other things, the size of a population of insects within an area.
  • One example embodiment system 10 is schematically shown in FIG. 1, and includes an optical insect detector subsystem 12. The optical insect detector subsystem 12 includes a plurality of light sources (or illuminators) 14 a,b and an equal number of phototransistors 16 a,b. The optical insect detector subsystem 12 is electrically connected to, and powered by a, a power source 18. The power source 18 may be, for example, a battery. In order to eliminate false readings, the illuminators 14 a,b should produce a constant brightness. Another important consideration, for best results, is that the illuminators 14 a,b (or the heat they produce) should not affect the behavior of the insects. Accordingly, a suitable light source is, for example, an LED that emits light at a wavelength of at, or around, 940 nm (infrared). This wavelength is not visible to insects (or humans) and does not produce any detectible heat.
  • A first illuminator 14 a and a first phototransistor 16 a face each other from some distance, such that some of the light 20 a that is emitted from the first illuminator 14 a falls upon the first phototransistor 16 a. In this embodiment, the optical insect detector subsystem 12 utilizes two illuminators 14 a,b and two corresponding phototransistors 16 a,b, which are placed in a parallel, side-by-side arrangement, some small distance apart, such that the light 20 a from the first illuminator 14 a only falls upon the first phototransistor 16 a, and light 20 b from a second illuminator 14 b falls only upon a second phototransistor 16 b. In this embodiment, the illuminators 14 a,b are placed approximately one centimeter apart. In the absence of insects, or anything else, that blocks part of the light 20 a,b, the amount of light 20 a,b that falls on the phototransistors 16 a,b is constant. Furthermore, by examining the timing difference between when an insect crosses the first light beam 20 a and the second light beam 20 b, the system 10 can infer the speed and direction of travel of the insect. The speed and direction of travel are two inputs that may be fed into the system's classification algorithms to identify the flying insect.
  • The phototransistors 16 a,b are connected to an analog to digital converter 22, and in a constant light field, no signal 23 (not shown) is produced. However, when an object blocks all or part of the light 20 a,b from the illuminators 14 a,b, the phototransistors 16 a,b will create a changing voltage that the analog to digital converter 22 will convert to a signal 23. The signal 23 produced by the analog to digital converter 22 is then captured by a recording device 24. While this signal 23 is recorded optically as described above, it is effectively a sound file, and can be saved in any standard computer sound file format (e.g., as a WAV file), and if played on a standard audio device sounds like a conventionally recorded sound file recorded with a microphone. Furthermore, since this embodiment utilizes two illuminators 14 a,b and two phototransistors 16 a,b, two separate signals 23 a,b will be produced by the analog to digital converter 22, which can be recorded as separate stereo tracks by the recording device 24.
  • The system 10 continuously records and monitors the two independent signals 23 a,b. In the absence of insects there should be no signal 23 produced. In practice, however, electronic devices always produce a tiny amount of signal due to noise in the circuit (for example, wires within the electronics may act as a weak antenna, picking up a tiny amount of radio signals or other electromagnetic interference). Accordingly, the system 10 will typically utilize a threshold to determine if the device is currently seeing a signal 23 caused by the presence of an insect or not. This threshold can be computed in several ways. For example, the device can be sealed so that there are no insects present, at which point the mean and standard deviation of the noise amplitude is measured. In one embodiment, the signal threshold is then set to be the measured mean plus four standard deviations.
  • In use, when an insect flies past the two light beams 20 a,b, the analog to digital converter 22 will produce signals 23 a,b whose amplitude greatly exceed the threshold. The recording device 24 then records a sound snippet of the interception event. In certain embodiments, the sound snippet is recorded from one second before the interception, to one second after the interception. The system 10 is able to record a sound snippet of this event one second before it happens, by maintaining the signals 23 a,b in a circular buffer. That is, the recording device 24 is constantly recording the signals 23 a,b in a constant loop, and when an interception event occurs, the recording is tagged to begin at one second prior in the recording loop.
  • The system 10 further utilizes sensors 26 to measure environmental conditions surrounding the system 10. Examples of information that may be measured by the sensors 26 include, but are not limited to, the humidity, temperature, air pressure, ambient light, date, time, and location on Earth. The sensors 26 are in electrical communication with the recording device 24, such that information measured by the sensors 26 is sent to be recorded by the recording device 24.
  • The recording device 24 is therefore able to create a holistic record of the insect interception event that may include, among other things, a stereo sound file containing the signals 23 a,b, assigned to specific stereo tracks; humidity; temperature; air pressure; ambient light; date; time; location on Earth, and/or the amplitude spectrum derived from the stereo sound file. The amplitude spectrum is a vector of real numbers, and provides an additional important piece of information useful in differentiating one species from another that have similar wingbeat frequencies. By additionally recording and analyzing the amplitude spectrum, error rates in correctly identifying the species are dramatically reduced. This is similar to the scenario wherein an individual can differentiate middle C played on a piano and middle C played on a violin. While middle C on both instruments has the same frequency, additional information is conveyed beyond merely the frequency. This additional information contained within the amplitude spectrum leads to a dramatic decrease in error rate from simply using the wingbeat frequency alone. In one embodiment, values that correspond to frequencies less than 20 Hz or greater than 2000 Hz are eliminated, since such data is very unlikely to contribute true signal generated by an insect. This information may be recorded in various sound file formats, for example WAV or MP3 formats. Furthermore, when recording the data, the two signals 23 a,b may be recorded as stereo sound tracks and the remaining data may be embedded within the sound file as metadata.
  • Offline Learning and Insect Models
  • While the identification system described above is greatly improved over that of the prior art, it still may face potential problems when attempting to identify unrelated flying insects that have similar wingbeat frequencies. However, with each new piece of information that can be provided to the classifier, the error rate is greatly reduced.
  • Any attempt to classify flying insects benefits from having high quality data models of the insect's flight behavior, and the prior probability of seeing that sex or species of insect at a given time, on a given day of the year at a given location on Earth. Moreover, it is known that insect's flight behavior depends on environmental conditions. However, under prior art practices, given all these variables, obtaining high quality data models that span the space of all possibilities has proven to be difficult, if not impossible. The methods of the present disclosure mitigate this problem with the addition of an offline learning step 28 and insect modeling steps 30 a,b.
  • This offline learning step 28 can be achieved by placing a system 10 into an insectary (not shown) with juvenile insects of a single species, and single sex and species for sexually dimorphic insects (i.e. pupa for mosquitoes, maggots for flies, bees and wasps, etc.) and any food, water or other items needed to sustain healthy adult insect. The system 10 is operated continuously to record the emerging insects, 24 hours a day, for their entire adult lives.
  • This recording is conducted at a fixed humidity, temperature, and air pressure, with a fixed artificial day and night cycle (e.g., 16 hours of darkness, followed by 8 hours of light, with a five minute linear ramp-up/ramp-down of the brightness to simulate dawn and dusk). A similar recording is conducted utilizing a different fixed artificial day and night cycle (e.g., 12 hours of darkness followed by 12 hours of light), with otherwise the same environmental conditions as the first recording.
  • Collecting data in the manner allows the creation of excellent insect models regarding the circadian rhythm 30 a of the insect, for those recorded environment conditions. However, the real world does not necessarily operate under such ideal conditions. As such, the methods of the present disclosure allow for the generalization from the recorded conditions in the offline learning step 28 to any reasonable single environment condition 30 b, as shown in the flowchart of FIG. 2.
  • Generalizing to Achieve Invariance to the Circadian Rhythm
  • The probability that a given insect will be flying at a given time of day (relative to sunrise/sunset) can be encoded in a flight activity circadian rhythm. Certain embodiments of systems and methods described herein may include a circadian rhythm subsystem 32 that can create an insect specific circadian rhythm for any day of the year, for any location on Earth, after seeing just two distinct example circadian rhythms for that insect in the offline learning step 28. Because this prior probability can vary by orders of magnitude over the course of the day, it greatly affects the ability to correctly identify insects.
  • For example, the daily flight rhythms of Culex tarsalis and Aedes aegypti are very different, and offer a useful feature for the Bayesian classifier used in one embodiment of the system 10. To show how this would be utilized, an insect detected at 3 am is about four times more likely to be a Cx. tarsalis, whereas an insect spotted at midday is almost certainly an Ae. aegypti.
  • The circadian rhythm subsystem 32, given two distinct circadian rhythms observed during the offline learning step 28, can produce a high quality approximation to the circadian rhythm that would be observed for any timing of dawn and dusk via the circadian rhythm modeling step 30 a. The system 10 creates a synthetic circadian rhythm in the circadian rhythm modeling step 30 a using a type of inverted Dynamic Time Warping (DTW) based on the information provided from the offline learning step 28. Normally, the DTW algorithm takes as input two time series, C and Q, and produces a warping path in an alignment matrix as a side effect of calculating the distance DTW(Q,C). In contrast, the system 10 uses a modified DTW algorithm to align the two given two distinct circadian rhythms to the desired circadian rhythm (the latter specified by desired time of dawn and dusk). The amount of x-axis “warping” needed to convert the observed circadian rhythms into the desired circadian rhythm is used as an input into a regression algorithm, which is then used to set the y-axis value of the circadian rhythm. Accordingly, the circadian rhythm subsystem 32 takes the synthetic circadian rhythm produced in step 30 a and the captured signals 23 a,b and feeds this information to the classifier 34.
  • Generalizing to Achieve Invariance to the Humidity Temperature and Air Pressure
  • Many research papers over the last four decades have examined how insect flight (usually, just considering the wingbeat frequency) varies with altitude, humidity, temperature or air pressure. However, all of the above mostly affect the flight behavior by affecting the density of the air. The density of the air for a given temperature or air pressure can be computed using the ideal gas law. Additionally, if it is desired to compensate for humidity (humid air is less dense than dry air) the density of humid air may be calculated as a mixture of ideal gases. As it is known that air density affects the amplitude spectrum, this can be adjusted to compensate for the differences in the ambient environment from ideal test conditions.
  • Given the above, in the offline learning step 28, data is collected about various insects at at least one canonical set of environmental conditions (typically, temp=80° F., Humidity=20%, light cycle 16|8). This data is then utilized in an environmental conditions modeling step 30 b to create an environmental conditions subsystem 36. Accordingly, the captured signals 23 a,b and information provided from the environmental sensors 26 can be input into the environmental conditions subsystem 36 to render data that has been compensated based on the environmental conditions to match those of ideal test conditions. Alternatively, one can collect data about various insects at ambient environmental conditions and normalize it to reflect what would have been observed at the canonical set of environmental conditions. The captured signals 23 a,b as modified by the environmental conditions subsystem 36 are then fed to the classifier 34.
  • Usage
  • The information gathered from the offline learning step 28, including amplitude spectrum, wingbeat frequency, daily flight rhythms, average flight speed and/or direction, etc. is used to build the classifier 34. While any classifier may be used, in certain embodiments a Bayesian Classifier is utilized.
  • During deployment, the system 10 is placed in the desired location and switched on. After the insect signals 23 a,b are captured, the system 10 normalizes the signals 23 a,b to achieve invariance to ambient altitude, humidity, temperature and air pressure with the environmental conditions subsystem 36, and to the current time of dawn/dusk with the circadian rhythm subsystem 32. The output from these subsystems 32, 36 is then input into the classifier 34, which produces the predicted class label of the sex and/or species and/or physiological state.
  • Given that the system 10 is able to classify the sex, species, and physiological states of insects in real time, the system 10 may optionally use the ability to classify a single insect, or the cumulative number/ratios of classified insects observed thus far, to control an actuator 38. For example, an actuator 38 can be used to selectively kill insects based on their sex. This is a potentially useful ability to support sterile insect technique (SIT) for mosquitoes. In such cases you want to release only males from a trap or insectary. However, other embodiments could allow for the actuator 38 to kill only a particular pest insect while allowing a beneficial insect to escape the trap unharmed. An exemplary kill method would utilize a powerful laser aimed at the exit of a trap or insectary, wherein the actuator 38 activates the laser to kill the desired insect. Upon attempting to exit the trap, the insect is classified via the system 10, and if it is determined to be a target insect, the actuator 38 activates the laser. When a non-target insect is identified, the system 10 does not activate the actuator 38.
  • In another embodiment, rather than killing a target insect, the actuator 38 may be utilized to capture the target insect. In this case, the system 10 is attached to an apparatus that can capture living insects, such as a simple plastic cup suspend over a flat surface. The cup may be held up by an electromagnet connected to the actuator 38. When a target insect is identified by the system 10, a signal is sent to the actuator 38 to turn off the electromagnet, causing the cup to drop on top of the target insect, thereby capturing it. Alternatively, the actuator 38 could close a door in a trap, thereby retaining the target insect inside the trap.
  • In yet another embodiment, the system 10 is used to control at least one valve solenoid. The solenoid may be opened by the actuator 38 to release chemicals in response to the identification of a single insect, or the cumulative number/ratios of identified insects observed thus far. Examples of chemicals that may be released include, but are not limited to, attractants, repellents, pesticides or larvicides.
  • Examples of commercial applications of the present systems and methods include:
  • Agricultural: Insect pests destroy billions of dollars worth of food each year. Any attempt to mitigate their harmful effects requires knowing the density of the species present (and sometimes the sex of the insects). The present systems can provide this species/sex count in real time.
  • Vector Control Districts: Most vector control districts spend considerable effort to quantify the density of various mosquitoes in their district. Using conventional methods, this information can be inaccurate, out-of-date (by days or even weeks) and expensive to obtain. The present systems can provide real-time accurate information at a very low amortized cost.
  • Residential Mosquito Control: There are several companies that sell devices to lure and kill mosquitoes. Such devices could be augmented with the present systems, to let the user know how successful the mosquito traps are in real time. Furthermore, this information could be actionable. For example, it could help the user optimize trap placement, or provide the user with information to reschedule an outdoor event if the noted density of mosquitoes is trending up.
  • The above description is given by way of example, and not limitation. Given the above disclosure, one skilled in the art could devise variations that are within the scope and spirit of the invention disclosed herein, including various ways of using the actuator 38 to control insect behavior in some way upon identifying a single insect or the cumulative number/ratios of identified insects, such as capturing the insect(s), killing the insect(s), and/or luring or repelling insect(s) with chemicals, sounds, or lights.
  • Further, the various features of the embodiments disclosed herein can be used alone, or in varying combinations with each other and are not intended to be limited to the specific combination described herein. Thus, the scope of the claims is not to be limited by the illustrated embodiments.

Claims (16)

What is claimed is:
1. A system for classifying flying insects, said system comprising:
a power source in electrical connection with an optical insect detector subsystem, said optical insect detector subsystem comprising:
a first illuminator that emits a first light transmission;
a first phototransistor positioned to receive the first light transmission;
a second illuminator that emits a second light transmission;
a second phototransistor positioned to receive the second light transmission;
a mechanical divider positioned between the first and second illuminators, such that the mechanical divider blocks the first light transmission from reaching the second phototransistor and blocks the second light transmission from reaching the first phototransistor; and
wherein the first and second illuminators are positioned a defined distance apart from each other;
an analog-to-digital converter in electrical communication with the first and second phototransistors, wherein the analog-to-digital converter is configured to produce a first digital signal from a change in voltage of the first phototransistor and to produce a second digital signal from a change in voltage of the second phototransistor;
a recording device configured to record the first and second digital signals; and
a classification subsystem configured to receive the first and second digital signals from the recording device and to output a flying insect identification based upon the first and second digital signals.
2. The system of claim 1, further comprising an actuation subsystem electronically connected to the classification subsystem and configured to actuate a device upon receiving an identification signal from the classification subsystem.
3. The system of claim 2, wherein the device is a laser configured to kill an insect based upon receiving the identification signal.
4. The system of claim 2, wherein the device is an electromagnetic capture device configured to capture an insect based upon receiving the identification signal.
5. The system of claim 4, wherein the capture device comprises a cup attached to an electromagnet.
6. The system of claim 4, wherein the capture device comprises an electromagnetically activated door.
7. The system of claim 2, wherein the device is a solenoid valve configured to open and release a chemical based upon receiving the identification signal.
8. The system of claim 7, wherein the chemical is selected from the group consisting of attractants, repellents, pesticides, and larvicides.
9. An apparatus for classifying flying insects, said apparatus comprising:
a power source in electrical connection with an optical insect detector unit, said optical insect detector unit comprising:
a first light emitting diode that emits a first light transmission having a wavelength of approximately 940 nm;
a first phototransistor positioned to receive the first light transmission;
a second light emitting diode that emits a second light transmission having a wavelength of approximately 940 nm;
a second phototransistor positioned to receive the second light transmission;
a mechanical divider positioned between the first and second light emitting diodes, such that the mechanical divider blocks the first light transmission from reaching the second phototransistor and blocks the second light transmission from reaching the first phototransistor; and
wherein the first and second light emitting diodes are positioned approximately one centimeter apart from each other;
an analog-to-digital converter in electrical communication with the first and second phototransistors, wherein the analog-to-digital converter is configured to produce a first digital signal from a change in voltage of the first phototransistor and to produce a second digital signal from a change in voltage of the second phototransistor;
a recording device configured to record the first and second digital signals as stereo tracks of an audio file;
an environmental sensor unit comprising:
at least one sensor configured to measure at least one environmental condition data point surrounding the system selected from the group consisting of the humidity, temperature, air pressure, ambient light, current date, current time, and current location,
wherein the environmental sensor unit is in electronic communication with the recording device, such that the recording device records the at least one environmental condition data point measured by the at least one sensor;
a circadian rhythm unit configured to calculate a synthetic circadian rhythm of an insect intended to be identified by the apparatus based upon at least one environmental condition data point measured by the environmental sensor unit; and
a classification unit configured to receive the first and second digital signals from the recording device and the at least one environmental condition data point and to output a flying insect identification based upon the first and second digital signals, at least one environmental condition data point, and the calculated synthetic circadian rhythm.
10. The apparatus of claim 9, wherein the apparatus is contained within or attached to an insect trap.
11. A method of classifying a flying insect comprising the following steps:
a) recording a first set of data of the life cycle of the flying insect at set environmental conditions utilizing a first day and night cycle, wherein the data recorded includes at least the insect's wingbeat frequency, amplitude spectrum, and circadian rhythm;
b) recording a second set of data of the life cycle of the flying insect at set environmental conditions utilizing a different, second day and night cycle wherein the data recorded includes at least the insect's wingbeat frequency, amplitude spectrum, and circadian rhythm;
c) creating a data model of the flying insect based upon the data recorded in steps (a) and (b);
d) inputting the data model created in step (c) into a classifier;
e) recording environmental conditions wherein the flying insect is to be classified;
f) creating a synthetic circadian rhythm based upon the data model created in step (c) and the recorded environmental conditions in step (e);
g) recording a first signal containing the wingbeat frequency and amplitude spectrum of the flying insect to be classified;
h) recording a second signal containing the wingbeat frequency and amplitude spectrum of the flying insect to be classified, wherein the second signal is captured at a known distance from the first signal;
i) processing the first and second signals recorded in steps (g) and (h) to normalize the signals based on the synthetic circadian rhythm created in step (f) and the environmental conditions recorded in step (e);
j) inputting the first and second signals processed in step (i) into the classifier;
k) classifying the insect based upon the signals inputted in step (j) and the data model input in step (d).
12. The method of claim 11, wherein the classifier is configured to classify the insect based on species.
13. The method of claim 11, wherein the classifier is configured to classify the insect based on sex.
14. The method of claim 11, wherein the classifier is configured to classify the insect based on wing variation.
15. The method of claim 11, wherein the environmental conditions recorded in step (e) are used to compensate for air density differences in step (i), and the resulting changes to the amplitude spectrum recorded in steps (g) and (h) from those recorded in steps (a) and (b) for the flying insect.
16. The method of claim 11, further comprising the step 1) operating an actuator based upon the classification result of step (k).
US17/366,993 2015-09-01 2021-07-02 Systems and methods for classifying flying insects Abandoned US20210329903A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US17/366,993 US20210329903A1 (en) 2015-09-01 2021-07-02 Systems and methods for classifying flying insects

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
US201562212744P 2015-09-01 2015-09-01
US15/253,286 US10178856B2 (en) 2015-09-01 2016-08-31 Systems and methods for classifying flying insects
US16/221,199 US20190110459A1 (en) 2015-09-01 2018-12-14 Systems and methods for classifying flying insects
US16/817,468 US20200205390A1 (en) 2015-09-01 2020-03-12 Systems and methods for classifying flying insects
US17/366,993 US20210329903A1 (en) 2015-09-01 2021-07-02 Systems and methods for classifying flying insects

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US16/817,468 Continuation US20200205390A1 (en) 2015-09-01 2020-03-12 Systems and methods for classifying flying insects

Publications (1)

Publication Number Publication Date
US20210329903A1 true US20210329903A1 (en) 2021-10-28

Family

ID=58103310

Family Applications (4)

Application Number Title Priority Date Filing Date
US15/253,286 Active 2037-07-18 US10178856B2 (en) 2015-09-01 2016-08-31 Systems and methods for classifying flying insects
US16/221,199 Abandoned US20190110459A1 (en) 2015-09-01 2018-12-14 Systems and methods for classifying flying insects
US16/817,468 Abandoned US20200205390A1 (en) 2015-09-01 2020-03-12 Systems and methods for classifying flying insects
US17/366,993 Abandoned US20210329903A1 (en) 2015-09-01 2021-07-02 Systems and methods for classifying flying insects

Family Applications Before (3)

Application Number Title Priority Date Filing Date
US15/253,286 Active 2037-07-18 US10178856B2 (en) 2015-09-01 2016-08-31 Systems and methods for classifying flying insects
US16/221,199 Abandoned US20190110459A1 (en) 2015-09-01 2018-12-14 Systems and methods for classifying flying insects
US16/817,468 Abandoned US20200205390A1 (en) 2015-09-01 2020-03-12 Systems and methods for classifying flying insects

Country Status (8)

Country Link
US (4) US10178856B2 (en)
EP (1) EP3345131B1 (en)
DK (1) DK3345131T3 (en)
ES (1) ES2883123T3 (en)
HU (1) HUE055037T2 (en)
PL (1) PL3345131T3 (en)
PT (1) PT3345131T (en)
WO (1) WO2017040814A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE102021133586A1 (en) 2021-12-17 2023-06-22 Fujin Beteiligungsgesellschaft Mbh Device and method for detecting and locating vermin
GB2622306A (en) * 2022-07-19 2024-03-13 Agrisound Ltd Apparatus and methods for generating a low resolution representation of an insect sound for classifying insect activity

Families Citing this family (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10178856B2 (en) * 2015-09-01 2019-01-15 Isca Technologies, Inc. Systems and methods for classifying flying insects
US20180303079A1 (en) * 2015-10-16 2018-10-25 The Trustees Of Columbia University In The City Of New York Acoustic Automated Detection, Tracking and Remediation of Pests and Disease Vectors
MX2019002644A (en) 2016-09-08 2019-06-17 Walmart Apollo Llc Systems and methods for dispensing an insecticide via unmanned vehicles to defend a crop-containing area against pests.
WO2020102744A1 (en) * 2018-11-16 2020-05-22 Verily Life Sciences Llc Systems and methods for sensing insect sex or species
CN109454027B (en) * 2018-12-14 2023-04-18 湖南元想科技有限公司 Insect classification detection equipment and method
EP3669648A1 (en) * 2018-12-18 2020-06-24 César Dos Santos, Gil Method of detection and elimination of non-endemic invasive species
US11006200B2 (en) * 2019-03-28 2021-05-11 Sonova Ag Context dependent tapping for hearing devices
EP3975710A1 (en) * 2019-05-24 2022-04-06 Romanova, Anastasiia Romanivna Mosquito monitoring and counting system
CN111830029B (en) * 2020-01-02 2023-10-20 河北盛鹏化工有限公司 Pesticide preparation concentration on-site analysis system and method
US20210368764A1 (en) * 2020-06-01 2021-12-02 Alliance Sports Group, L.P. Multi-Frequency Targeting Insect Control
US11490609B2 (en) * 2020-06-25 2022-11-08 Satish K. CHerukumalli Mosquito identification classification trap and method to use
EP4033409A1 (en) 2021-01-20 2022-07-27 Smart Farm Sensing B.V. Insect trap detection method, sensor device and system
CN114279475B (en) * 2021-12-23 2023-12-26 成都德鲁伊科技有限公司 Biological behavior monitoring method by utilizing light sensor
US20230341583A1 (en) * 2022-04-25 2023-10-26 FarmSense, Inc. Pest detection systems and methods
CN115362990B (en) * 2022-09-22 2024-02-13 中国农业科学院植物保护研究所 Insect wing vibration frequency measuring device and method

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140226860A1 (en) * 2009-01-15 2014-08-14 Tokitae Llc Photonic fence
US10178856B2 (en) * 2015-09-01 2019-01-15 Isca Technologies, Inc. Systems and methods for classifying flying insects

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6653971B1 (en) * 1999-05-14 2003-11-25 David L. Guice Airborne biota monitoring and control system
US7496228B2 (en) 2003-06-13 2009-02-24 Landwehr Val R Method and system for detecting and classifying objects in images, such as insects and other arthropods
WO2005098379A1 (en) 2004-04-08 2005-10-20 Purdue Research Foundation Multi-spectral detector and analysis system
CA2646984C (en) 2006-03-21 2015-07-14 Board Of Regents, The University Of Texas System Optical device for detecting live insect infestation
PL2052236T3 (en) * 2006-08-01 2012-10-31 Photonic Detection Systems Pty Ltd Optical sensing system and method for discriminating plant matter
CN102340988B (en) 2009-01-15 2014-02-26 脱其泰有限责任公司 Photonic fence
TWI442884B (en) 2011-09-29 2014-07-01 Nat Univ Tsing Hua System for classifying tiny insects
EP3107382B1 (en) * 2014-02-18 2020-12-02 onVector Technology LLC Object detection systems

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140226860A1 (en) * 2009-01-15 2014-08-14 Tokitae Llc Photonic fence
US10178856B2 (en) * 2015-09-01 2019-01-15 Isca Technologies, Inc. Systems and methods for classifying flying insects

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE102021133586A1 (en) 2021-12-17 2023-06-22 Fujin Beteiligungsgesellschaft Mbh Device and method for detecting and locating vermin
GB2622306A (en) * 2022-07-19 2024-03-13 Agrisound Ltd Apparatus and methods for generating a low resolution representation of an insect sound for classifying insect activity

Also Published As

Publication number Publication date
US20190110459A1 (en) 2019-04-18
US10178856B2 (en) 2019-01-15
EP3345131A4 (en) 2019-04-03
HUE055037T2 (en) 2021-10-28
WO2017040814A1 (en) 2017-03-09
US20200205390A1 (en) 2020-07-02
ES2883123T3 (en) 2021-12-07
DK3345131T3 (en) 2021-07-26
PL3345131T3 (en) 2021-11-08
PT3345131T (en) 2021-07-16
US20170055511A1 (en) 2017-03-02
EP3345131A1 (en) 2018-07-11
EP3345131B1 (en) 2021-04-14

Similar Documents

Publication Publication Date Title
US20210329903A1 (en) Systems and methods for classifying flying insects
Batista et al. Towards automatic classification on flying insects using inexpensive sensors
US11771074B2 (en) Sensor based observation of anthropods
Potamitis et al. Affordable bimodal optical sensors to spread the use of automated insect monitoring
EP3107382B1 (en) Object detection systems
Potamitis et al. Novel noise-robust optoacoustic sensors to identify insects through wingbeats
Chen et al. Flying insect classification with inexpensive sensors
Silva et al. Applying machine learning and audio analysis techniques to insect recognition in intelligent traps
US10914837B2 (en) Object detection systems
US11523599B2 (en) Arthropod detection
Batista et al. SIGKDD demo: sensors and software to allow computational entomology, an emerging application of data mining
Potamitis et al. Large aperture optoelectronic devices to record and time-stamp insects’ wingbeats
CN106793768A (en) Photonic fence
de Souza et al. Classification of data streams applied to insect recognition: Initial results
Greif et al. Using on-board sound recordings to infer behaviour of free-moving wild animals
Schmieder et al. Sensory constraints on prey detection performance in an ensemble of vespertilionid understorey rain forest bats
Santos et al. Automated electronic approaches for detecting disease vectors mosquitoes through the wing-beat frequency
BR112019006453A2 (en) method, computer program product and system
González-Pérez et al. A novel optical sensor system for the automatic classification of mosquitoes by genus and sex with high levels of accuracy
van Roy et al. Optical identification of bumblebee species: effect of morphology on wingbeat frequency
Kim et al. Infrared light sensors permit rapid recording of wingbeat frequency and bioacoustic species identification of mosquitoes
Kalfas et al. Towards in-field insect monitoring based on wingbeat signals: The importance of practice oriented validation strategies
Symes et al. From understory to canopy: In situ behavior of Neotropical forest katydids in response to bat echolocation calls
Banlawe et al. Decision tree learning algorithm and naïve Bayes classifier algorithm comparative classification for mango pulp weevil mating activity
Murugeshan Acoustic monitoring of airborne insects in outdoor environments

Legal Events

Date Code Title Description
AS Assignment

Owner name: ISCA TECHNOLOGIES, INC, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MAFRA-NETO, AGENOR;REEL/FRAME:056758/0165

Effective date: 20160831

Owner name: THE REGENTS OF THE UNIVERSITY OF CALIFORNIA, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:KEOGH, EAMONN;REEL/FRAME:056747/0403

Effective date: 20160925

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION