US12298245B2 - Detection device, assistance device and assistance method - Google Patents

Detection device, assistance device and assistance method Download PDF

Info

Publication number
US12298245B2
US12298245B2 US17/859,364 US202217859364A US12298245B2 US 12298245 B2 US12298245 B2 US 12298245B2 US 202217859364 A US202217859364 A US 202217859364A US 12298245 B2 US12298245 B2 US 12298245B2
Authority
US
United States
Prior art keywords
detection
detection portion
position information
substance
plant
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US17/859,364
Other versions
US20230096759A1 (en
Inventor
Katsumi Maeda
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
NEC Corp
Original Assignee
NEC Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by NEC Corp filed Critical NEC Corp
Assigned to NEC CORPORATION reassignment NEC CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MAEDA, KATSUMI
Publication of US20230096759A1 publication Critical patent/US20230096759A1/en
Application granted granted Critical
Publication of US12298245B2 publication Critical patent/US12298245B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/62Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light
    • G01N21/63Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light optically excited
    • G01N21/64Fluorescence; Phosphorescence
    • G01N21/645Specially adapted constructive features of fluorimeters
    • G01N21/6456Spatial resolved fluorescence measurements; Imaging
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/62Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light
    • G01N21/63Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light optically excited
    • G01N21/64Fluorescence; Phosphorescence
    • G01N21/6428Measuring fluorescence of fluorescent products of reactions or of fluorochrome labelled reactive substances, e.g. measuring quenching effects, using measuring "optrodes"
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/75Systems in which material is subjected to a chemical reaction, the progress or the result of the reaction being investigated
    • G01N21/77Systems in which material is subjected to a chemical reaction, the progress or the result of the reaction being investigated by observing the effect on a chemical indicator
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N33/00Investigating or analysing materials by specific methods not covered by groups G01N1/00 - G01N31/00
    • G01N33/0098Plants or trees
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/22Image preprocessing by selection of a specific region containing or referencing a pattern; Locating or processing of specific regions to guide the detection or recognition
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • G06V20/188Vegetation
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/75Systems in which material is subjected to a chemical reaction, the progress or the result of the reaction being investigated
    • G01N21/77Systems in which material is subjected to a chemical reaction, the progress or the result of the reaction being investigated by observing the effect on a chemical indicator
    • G01N2021/7769Measurement method of reaction-produced change in sensor
    • G01N2021/7786Fluorescence
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N2201/00Features of devices classified in G01N21/00
    • G01N2201/02Mechanical
    • G01N2201/022Casings
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10064Fluorescence image
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30181Earth observation
    • G06T2207/30188Vegetation; Agriculture

Definitions

  • the present disclosure relates to a method for detecting plant diseases.
  • a pathological diagnosis system for plants described in International Publication No. WO 2021/111621 performs pathology determination based on a spot of a crop in an image captured by a camera mounted on a drone.
  • An object of the present disclosure is to provide a device or the like enabling recognition of the presence of a plant with an early-stage disease.
  • a detection device includes a detection portion that contains a substance that emits fluorescence in a case of being irradiated with light when reacting with a plant hormone.
  • An assistance device includes: a generation unit that generates position information of a plant infected with a disease based on a captured image of a detection portion irradiated with light and a position of the detection portion, the detection portion containing a substance that emits fluorescence in a case of being irradiated with the light when reacting with a plant hormone; and an output unit that outputs the generated position information.
  • a terminal device includes: a receiving unit that receives position information of a plant infected with a disease, the position information being generated based on a captured image of a detection portion irradiated with light and a position of the detection portion, and the detection portion containing a substance that emits fluorescence in a case of being irradiated with the light when reacting with a plant hormone; and a display unit that displays the received position information.
  • An assistance method includes: generating position information of a plant infected with a disease based on a captured image of a detection portion irradiated with light and a position of the detection portion, the detection portion containing a substance that emits fluorescence in a case of being irradiated with the light when reacting with a plant hormone; and outputting the generated position information.
  • a non-transitory computer-readable storage medium stores a program for causing a computer to execute: generation processing of generating position information of a plant infected with a disease based on a captured image of a detection portion irradiated with light and a position of the detection portion, the detection portion containing a substance that emits fluorescence in a case of being irradiated with the light when reacting with a plant hormone; and output processing of outputting the generated position information.
  • FIG. 1 is a schematic diagram schematically illustrating an example of a configuration of a detection device according to a first example embodiment of the present disclosure
  • FIG. 2 is a schematic diagram schematically illustrating an example of a configuration of a detection device according to a second example embodiment of the present disclosure
  • FIG. 3 is a schematic diagram schematically illustrating an example of a configuration of a detection device according to a third example embodiment of the present disclosure
  • FIG. 4 is a block diagram illustrating an example of a configuration of an assistance device according to a fourth example embodiment of the present disclosure
  • FIG. 5 is a flowchart illustrating an example of an operation of the assistance device according to the fourth example embodiment of the present disclosure
  • FIG. 6 is a block diagram illustrating an example of a configuration of an assistance system according to a fifth example embodiment of the present disclosure
  • FIG. 7 is a block diagram illustrating an example of an operation of the assistance device according to the fifth example embodiment of the present disclosure.
  • FIG. 8 is a block diagram illustrating an example of a configuration of a terminal device according to a sixth example embodiment of the present disclosure.
  • FIG. 9 is a diagram illustrating an example of a hardware configuration of a computer that can implement the assistance devices according to the above-described example embodiments.
  • FIG. 1 is a schematic diagram schematically illustrating an example of a configuration of a detection device according to a first example embodiment of the present disclosure.
  • the detection device 100 includes a detection portion 110 .
  • the detection portion 110 shows a reaction that is in proportion to the amount of infection signal substance of a disease of a plant. Specifically, when the detection portion 110 is irradiated with light, the detection portion 110 irradiated with light emits fluorescence with an intensity that is in proportion to the amount of the infection signal substance that has reacted with the detection portion 110 .
  • the infection signal substance is also referred to as a plant hormone.
  • the detection portion 110 contains, for example, a substance whose property is changed due to the infection signal substance of a disease of a plant.
  • the infection signal substance is, for example, a substance released from a plant infected with a disease, such as methyl salicylate or methyl jasmonate.
  • the plant disease of the present disclosure is a disease that causes an infected plant to release the infection signal substance.
  • a plant is infected with a disease (specifically, in a case where a plant is infected with pathogenic bacteria or the like)
  • the plant begins to release the infection signal substance before a visible symptom (for example, a spot caused by a disease of a leaf of the plant, and the like) appears in the plant.
  • a visible symptom for example, a spot caused by a disease of a leaf of the plant, and the like
  • the plant infected with the disease releases the infection signal substance without having the symptom, and then, the symptom appears in the plant (that is, a symptomatic state). Therefore, the present example embodiment has an effect that a plant infected with a disease can be detected before a visible symptom appears in the plant due to the disease.
  • the substance contained in the detection portion 110 is, for example, at least one of the following substances.
  • a rare earth and a zinc compound emit fluorescence in a case of being irradiated with ultraviolet light having a wavelength of 200 to 400 nanometers (nm).
  • a hydrazine derivative emits fluorescence in a case of being irradiated with ultraviolet light or visible light having a wavelength of 250 to 500 nm.
  • the light with which the detection portion 110 is irradiated is light including ultraviolet light having a wavelength of 200 to 400 nm.
  • the light with which the detection portion 110 is irradiated is light including at least one of ultraviolet light or visible light having a wavelength of 250 to 500 nm.
  • a first substance is a boron-oxygen compound having a structure represented by the following General Formula (1) as a receptor that selectively recognizes methyl salicylate.
  • the boron-oxygen compound may be a diboronic acid derivative represented by the following General Formula (2).
  • R represents a methyl group, an ethyl group, a propyl group, a butyl group, or a phenyl group, and each R may be the same or different.
  • the boron-oxygen compound may be a phenylboronic acid derivative represented by the following General Formula (3).
  • X represents a hydrogen molecule, an alkyl group having 1 to 4 carbon atoms, an alkoxy group, a mercapto group, a mercapto-substituted alkyl group, or a mercapto-substituted alkoxy group
  • n represents an integer of 1 to 5, and when n is equal to or more than 2, each X may be the same or different.
  • the boron-oxygen compound may be a boroxine derivative represented by the following General Formula (4).
  • Z represents an alkyl group having 1 to 4 carbon atoms or a substituted or unsubstituted phenyl group, and each Z may be the same or different.
  • the first substance emits fluorescence as methyl salicylate reacts with the boron-oxygen compound to form a boron complex.
  • a second substance is a substance in which the receptor that selectively recognizes methyl salicylate is a rare earth compound.
  • the above-described rare earth compound may be a compound of any of an acetate, a chloride, an oxalate, a nitrate, a propionate, an isobutyrate, and a pivalate of scandium (Sc), yttrium (Y), lanthanum (La), cerium (Ce), praseodymium (Pr), neodymium (Nd), promethium (Pm), samarium (Sm), europium (Eu), gadolinium (Gd), terbium (Tb), dysprosium (Dy), holmium (Ho), erbium (Er), thulium (Tm), ytterbium (Yb), and lutetium (Lu).
  • the above-described rare earth compound may be a compound of any of an acetate, a chloride, an oxalate, a nitrate, a propionate, an isobutyrate, or a pivalate of samarium (Sm), europium (Eu), gadolinium (Gd), terbium (Tb), or dysprosium (Dy).
  • the above-described rare earth compound may be a compound in which a salt of a rare earth element forms a complex with a phosphine oxide derivative.
  • the above-described rare earth compound emits fluorescence as methyl salicylate reacts with the rare earth compound to form a complex of rare earth.
  • a third substance is a substance in which the receptor that selectively recognizes methyl salicylate is a zinc compound.
  • the above-described zinc compound is a compound of any of zinc(II) acetate, zinc(II) formate dihydrate, zinc(II) butyrate, zinc oxalate dihydrate, zinc(II) hexanoate, zinc(II) propionate, zinc(II) tartrate dihydrate, zinc(II) benzoate, zinc(II) octanoate, zinc(II) oleate, zinc(II) nitrate, and zinc(II) chloride.
  • the above-described zinc compound emits fluorescence as methyl salicylate reacts with the zinc compound to form a complex.
  • a fourth substance is a substance in which the receptor that selectively recognizes methyl jasmonate is a hydrazine derivative.
  • the above-described hydrazine derivative is a compound of any of 4-hydrazino-7-nitro-2,1,3-benzoxadiazole hydrazine, 5-dimethylaminonaphthalene-1-sulfonyl hydrazine, and 4-(N,N-dimethylaminosulfonyl)-7-hydrazino-2,1,3-benzoxadiazole.
  • the above-described hydrazine derivative emits fluorescence as methyl jasmonate reacts with the hydrazine derivative to form a hydrazone derivative.
  • the detection portion 110 contains at least one of these substances.
  • the present example embodiment has an effect of enabling recognition of the presence of a plant with an early-stage disease. This is because, in a case of being irradiated with light, the detection portion 110 emits fluorescence with an intensity based on the amount of the infection signal substance of a disease of a plant that has reacted with the detection portion 110 per unit area.
  • the detection device 100 can be implemented by, for example, a base such as a plate to which the detection portion 110 such as a filter paper impregnated with the substance is attached.
  • the material of the base may be a combination of one or more of wood, metal, resin, mineral, and the like.
  • the material of the base is not particularly limited.
  • the shape of the base is not particularly limited.
  • the base may be formed as, for example, a plate, a column, a thread, a string, a cable, a tape, a sheet, or the like.
  • the detection portion 110 may be, for example, an object impregnated with the substance.
  • the detection portion 110 may be, for example, a filter paper, a cloth, a resin sheet that can be impregnated with the substance, or the like.
  • the detection portion 110 may be the substance applied onto the base or a mixture of the substance and another substance applied onto the base.
  • the shape of the detection portion 110 is not particularly limited.
  • the shape of the detection portion 110 may be a predetermined figure such as a rectangle, a polygon, or a circle.
  • the shape of the detection portion 110 may be a contour line of a predetermined figure having a finite thickness.
  • the shape of the detection portion 110 may be a mark, a character, a character string, or the like.
  • the shape of the detection portion 110 may be a graphic such as a barcode or a quick response (QR) code (registered trademark).
  • the detection device 100 may be a resin substrate, a wood substrate, a metal substrate, or the like to which a sheet such as a filter paper impregnated with the substance is attached.
  • the detection device 100 may be a substrate onto which a substance containing the substance is applied.
  • the detection device 100 may be a part of a structure such as a garden stone or a wall to which a sheet such as a filter paper impregnated with the substance is attached.
  • the detection device 100 may be a part of a structure such as a garden stone or a wall onto which a substance including the substance is applied.
  • the detection device 100 may be a tape onto which the substance is applied.
  • the material of the tape is, for example, paper, resin, a combination of paper and resin, or another material that can form the tape.
  • the detection device 100 may be a tape impregnated with the substance.
  • the detection device 100 may be a thread, a string, a cable, an optical fiber, or the like onto which the substance is applied.
  • the optical fiber is an optical fiber configured in such a way that a part of light passing through the inside leaks from the surface.
  • the detection device 100 may be a thread or a string impregnated with the substance.
  • the detection device 100 may be a cable or an optical fiber coated with a substance impregnated with the substance. In a case where the detection device 100 has a portion impregnated with the substance, the portion impregnated with the substance functions as the detection portion 110 . In a case where the detection device 100 has a portion onto which the substance is applied, the portion onto which the substance is applied functions as the detection portion 110 .
  • the outline of the detection portion 110 may be drawn by using a fluorescent substance that emits fluorescence when irradiated with light.
  • FIG. 2 is a schematic diagram schematically illustrating an example of a configuration of a detection device according to a second example embodiment of the present disclosure.
  • a detection device 101 includes a detection portion 110 , a reference portion 120 , and a shielding portion 130 .
  • the detection portion 110 and the reference portion 120 of the present example embodiment are the same as the detection portion 110 of the first example embodiment.
  • the shielding portion 130 is a detachable component that shields the reference portion 120 from the outside air containing the infection signal substance. Specifically, the shielding portion 130 covers the reference portion 120 in such a way that the reference portion 120 is shielded from the outside air except when the shielding portion 130 is removed from the reference portion 120 .
  • the shielding portion 130 can suppress a reaction between the substance in the reference portion 120 and the infection signal substance contained in the outside air by covering the reference portion 120 .
  • the outside air is air other than air between the shielding portion 130 and the reference portion 120 .
  • the present example embodiment has the same effects as those of the first example embodiment.
  • the reason is the same as the reason why the effects of the first example embodiment are exhibited.
  • the detection device 101 according to the second example embodiment of the present disclosure has a configuration similar to that of the detection device 100 according to the first example embodiment.
  • a portion that is not shielded by the shielding portion 130 may function as the detection portion 110
  • a portion that is shielded by the shielding portion 130 may function as the reference portion 120 .
  • a portion that is not shielded by the shielding portion 130 may function as the detection portion 110
  • a portion that is shielded by the shielding portion 130 may function as the reference portion 120 .
  • the detection portion 110 and the reference portion 120 may be formed to be separated from each other.
  • the detection portion 110 and the reference portion 120 may be integrally formed.
  • the shielding portion 130 is a component that covers the reference portion 120 in such a way that the reference portion 120 is shielded from the outside air, and is detachable from the reference portion 120 .
  • the material and shape of the shielding portion 130 are not particularly limited.
  • the outline of the detection portion 110 may be drawn by using a fluorescent substance that emits fluorescence when irradiated with light.
  • the outline of the reference portion 120 may be drawn by using a fluorescent substance that emits fluorescence when irradiated with light.
  • FIG. 3 is a schematic diagram schematically illustrating an example of a configuration of a detection device 102 according to a third example embodiment of the present disclosure.
  • the detection device 102 includes a detection portion 110 and an identification information portion 140 .
  • the detection device 102 may further include a reference portion 120 and a shielding portion 130 .
  • the detection portion 110 is the same as the detection portions 110 of the first example embodiment and the second example embodiment.
  • the reference portion 120 and the shielding portion 130 are the same as the reference portion 120 and the shielding portion 130 of the second example embodiment.
  • the identification information portion 140 holds identification information assigned to the detection device 102 in such a way that the identification information can be read from the outside of the detection device 102 .
  • the identification information is information that is assigned to the detection device 102 and can individually specify the detection device 102 .
  • the identification information may be, for example, a number such as a serial number.
  • the identification information may be represented by a character string.
  • the identification information may be represented by data.
  • the identification information may be represented by a barcode, a QR code, or the like.
  • the present example embodiment has the same effects as those of the first example embodiment.
  • the reason is the same as the reason why the effects of the first example embodiment are exhibited.
  • the detection portion 110 of the present example embodiment can be implemented by the specific example of the detection portion 110 described as the specific example of the first example embodiment.
  • the reference portion 120 and the shielding portion 130 of the present example embodiment can be implemented by the reference portion 120 and the shielding portion 130 described as the specific example of the second example embodiment.
  • the identification information portion 140 may be a portion of the surface of the detection device 102 on which the identification information is drawn.
  • the identification information is, for example, a number or another character string, a figure, a barcode, a QR code, or the like.
  • the identification information may be drawn by using a fluorescent substance.
  • the identification information may be drawn by using the substance.
  • the identification information portion 140 may also function as the detection portion 110 .
  • the identification information portion 140 may be a radio frequency identifier (RFID) tag holding the identification information.
  • RFID radio frequency identifier
  • FIG. 4 is a block diagram illustrating an example of a configuration of an assistance device 200 according to the fourth example embodiment of the present disclosure.
  • the assistance device 200 includes a generation unit 210 and an output unit 220 .
  • the generation unit 210 generates position information of a plant infected with a disease based on a captured image of the detection portion 110 irradiated with light and a position of the detection portion 110 , the detection portion 110 containing a substance that emits fluorescence in a case of being irradiated with the light when reacting with a plant hormone.
  • the detection portion is the detection portion 110 according to any one of the first to third example embodiments described above.
  • the output unit 220 outputs the generated position information.
  • the generation unit 210 may estimate information on the amount of the plant hormone at the position of the detection portion 110 from the brightness of the detection portion 110 in the captured image based on a relationship between the brightness of the detection portion 110 irradiated with ultraviolet light and the amount of the plant hormone that has reacted with the detection portion 110 .
  • the amount of the plant hormone that has reacted with the detection portion 110 may be, for example, the amount of the plant hormone that has reacted with the detection portion 110 per unit area.
  • the position of the detection device 100 including the detection portion 110 is regarded as the position of the detection portion 110 .
  • the amounts of the plant hormone at a plurality of time points are expressed as the trend of the amount of the plant hormone.
  • the generation unit 210 specifies the detection portion 110 (that is, the detection device 100 ) around which a plant infected with a disease is likely to exist based on the trend of the amount of the plant hormone that has reacted with the detection portion 110 .
  • the generation unit 210 generates the position information of the plant infected with the disease based on the specified position of the detection portion 110 .
  • the generation unit 210 may specify the detection portion 110 as the detection portion 110 around which a plant infected with a disease is likely to exist.
  • the generation unit 210 may specify the detection portion 110 as the detection portion 110 around which a plant infected with a disease is likely to exist.
  • the detection portion 110 around which a plant infected with a disease is likely to exist is also referred to as a detection receptor.
  • the generation unit 210 generates a range based on the position of the detection receptor as information indicating a region where a plant infected with a disease is likely to exist, in other words, the position information of the plant infected with the disease.
  • the generation unit 210 may generate information indicating the position of the detection receptor (for example, the position of the detection device 100 including the detection receptor) as the position information of the plant infected with the disease. Specifically, the generation unit 210 generates, as the position information of the plant infected with the disease, information of a screen in which a mark (for example, a combination of one or more of a figure, a symbol, a character, a character string, and the like) indicating the position of the detection receptor is superimposed on an image indicating a place in which the plant is cultivated.
  • a mark for example, a combination of one or more of a figure, a symbol, a character, a character string, and the like
  • the generation unit 210 may specify a region based on a distance from the position of the detection receptor (for example, a place in which the detection device 100 including the detection receptor is installed) as the region where a plant infected with a disease is likely to exist.
  • the region based on the distance from the position of the detection receptor may be, for example, a region within a predetermined distance from the position of the detection receptor.
  • the region based on the distance from the position of the detection receptor may be, for example, a region that is at a first predetermined distance or more from the position of the detection receptor and is at a second predetermined distance or less from the position of the detection receptor, the second predetermined distance being larger than the first predetermined distance.
  • the generation unit 210 may estimate a distance from the detection portion 110 to the plant infected with the disease based on the amount of change in amount of the plant hormone in the trend of the amount of the plant hormone that has reacted with the detection portion 110 .
  • the generation unit 210 holds in advance a relationship between the amount of change in amount of the plant hormone and the distance from the detection portion to the plant infected with the disease, and estimates the distance from the detection portion 110 to the plant infected with the disease based on the relationship.
  • the generation unit 210 may estimate the distance from the detection portion 110 to the plant infected with the disease based on an increment of the change in amount of the plant hormone in the trend of the amount of the plant hormone that has reacted with the detection portion 110 .
  • the generation unit 210 holds in advance a relationship between the increment of the change in amount of the plant hormone and the distance from the detection portion to the plant infected with the disease, and estimates the distance from the detection portion 110 to the plant infected with the disease based on the relationship.
  • the region based on the distance from the position of the detection receptor may be, for example, a region within the estimated distance from the position of the detection receptor.
  • the region based on the distance from the position of the detection receptor may be, for example, a region that is at a third predetermined distance or more from the position of the detection receptor and is at a fourth predetermined distance or less from the position of the detection receptor.
  • the third predetermined distance is smaller than the estimated distance
  • the fourth predetermined distance is larger than the estimated distance.
  • the generation unit 210 may specify, among a plurality of divided sections of the place in which the plant is cultivated, a section including a region within a predetermined distance from a place in which the detection device 100 including the detection receptor is installed, as the region where a plant infected with a disease is likely to exist.
  • the generation unit 210 generates, as the position information of the plant infected with the disease, information indicating a region where the plant infected with the disease is likely to exist. Specifically, the generation unit 210 generates information of a screen in which a mark indicating the region where the plant infected with the disease is likely to exist is superimposed on an image indicating the place in which the plant is cultivated.
  • the mark indicating the region where the plant infected with the disease is likely to exist may be a line surrounding the region where the plant infected with the disease is likely to exist.
  • the mark indicating the region where the plant infected with the disease is likely to exist may be a predetermined pattern drawn on the region where the plant infected with the disease is likely to exist.
  • the mark indicating the region where the plant infected with the disease is likely to exist may be changing the color of the region where the plant infected with the disease is likely to exist.
  • the mark indicating the region where the plant infected with the disease is likely to exist is not limited to these examples.
  • the position information of the infected plant is not limited to the above example as long as it is information indicating a position, a figure, a range, and the like, for example.
  • the position information of the infected plant may be text data or the like.
  • a place (and a region) where a plant is cultivated may be the inside of a greenhouse.
  • the place in which the plant is cultivated may be outdoors.
  • FIG. 5 is a flowchart illustrating an example of an operation of the assistance device 200 according to the fourth example embodiment of the present disclosure.
  • the generation unit 210 generates the position information of a plant infected with a disease based on a captured image of the detection portion 110 and the position of the detection portion 110 (Step S 101 ).
  • the output unit 220 outputs the position information of the plant infected with the disease (Step S 102 ).
  • the present example embodiment has an effect of enabling recognition of the presence of a plant with an early-stage disease. This is because the generation unit 210 generates the position information of the plant infected with the disease based on the captured image of the detection portion 110 and the position of the detection portion 110 . In a case of being irradiated with light, the detection portion 110 emits fluorescence with an intensity based on the amount of the infection signal substance of the disease of the plant that has reacted with the detection portion 110 per unit area.
  • FIG. 6 is a block diagram illustrating an example of a configuration of an assistance system according to a fifth example embodiment of the present disclosure.
  • an assistance system 1 includes a plurality of detection devices 100 , an assistance device 201 , a light source 300 , a sensor 400 , and a terminal device 500 .
  • the assistance device 201 is communicably connected to each of the sensor 400 and the terminal device 500 .
  • the light source 300 emits light.
  • the light source 300 is disposed in such a way as to irradiate a detection portion 110 of the detection device 100 with light.
  • the sensor 400 is, for example, a sensor that measures the brightness of the detection portion 110 of the detection device 100 irradiated with light from the light source 300 , obtained by imaging.
  • the sensor 400 is, for example, an imaging device that captures an image of the detection device 100 irradiated with light from the light source 300 (specifically, an image of at least a part of the detection portion 110 irradiated with light).
  • the imaging device is, for example, a digital camera, a camera of a smartphone, or the like.
  • the sensor 400 has a function of transmitting a measurement result (for example, an image) to the assistance device 201 .
  • the image is a moving image or one or more still images. For example, a worker performs measurement on the detection device 100 by using the sensor 400 .
  • the sensor 400 may be disposed in such a way as to capture an image of the detection device 100 including at least a part of a region of the detection portion 110 .
  • the sensor 400 may be disposed in such a way as to capture an image including only the region of the detection portion 110 . Note that parameters that affect a relationship between the brightness and color of the detection portion 110 and the brightness and color of the detection portion 110 in the image, such as the aperture, sensitivity, gain, and color temperature of the camera of the sensor 400 , are fixed.
  • the assistance device 201 includes a measurement data acquisition unit 230 , a position information acquisition unit 240 , a detection unit 250 , a measurement data storage unit 260 , a generation unit 210 , and an output unit 220 .
  • the generation unit 210 and the output unit 220 of the present example embodiment are the same as the generation unit 210 and the output unit 220 of the fourth example embodiment except for differences described below.
  • the measurement data acquisition unit 230 acquires measurement data from the sensor 400 .
  • the measurement data is, for example, an image of the detection portion 110 of the detection device 100 obtained by the sensor 400 .
  • the measurement data acquisition unit 230 transmits the acquired measurement data (for example, an image) to the detection unit 250 .
  • the measurement data acquisition unit 230 of the present example embodiment may acquire the measurement data of each of the plurality of detection portions 110 .
  • the position information acquisition unit 240 acquires the position of the detection portion 110 .
  • the position information acquisition unit 240 may acquire the position of the detection device 100 as the position of the detection portion 110 .
  • the position information acquisition unit 240 acquires the position information of each of the plurality of detection portions 110 .
  • the position information acquisition unit 240 may acquire, via an input device such as a keyboard or a touch panel, information indicating the position of the detection portion 110 , the information being input by an operator of the assistance system 1 using the input device.
  • the information indicating the position of the detection portion 110 may be, for example, a number assigned to the detection portion 110 .
  • the information indicating the position of the detection portion 110 may be, for example, a number assigned to the detection device 100 including the detection portion 110 . In these cases, a combination of a number that is information indicating a position and position information to which the number is assigned is assigned in advance to the position information acquisition unit 240 .
  • the position information acquisition unit 240 specifies, as the position information of the detection portion 110 , position information to which a number indicating the acquired position is assigned.
  • the position information may be identification information of a section of a plant cultivation field.
  • the position information may be coordinates in a coordinate system defined in the plant cultivation field.
  • the input information indicating the position of the detection portion 110 may be the position information of the detection portion 110 such as the identification information of the section or the coordinates.
  • the position information acquisition unit 240 transmits the position information of the detection portion 110 to the detection unit 250 .
  • the position information acquisition unit 240 may transmit, as the position information of the detection portion 110 , the position information of the detection device 100 to the detection unit 250 .
  • the detection unit 250 receives the measurement data (for example, an image) from the measurement data acquisition unit 230 . Furthermore, the detection unit 250 receives the position information of the detection portion 110 from the position information acquisition unit 240 .
  • the detection unit 250 detects a region of the detection portion 110 in the received image.
  • the detection unit 250 detects the entire image as the region of the detection portion 110 .
  • the detection unit 250 may calculate the intensity that is in proportion to the intensity of a wavelength (for example, 545 to 546 nm) of fluorescence, from a pixel value of each pixel of the image that is the measurement data.
  • An image in which a pixel shows the intensity of the wavelength of fluorescence is referred to as a fluorescence image.
  • the detection unit 250 extracts the region of the detection portion 110 in the fluorescence image based on the shape of the detection portion 110 .
  • the detection unit 250 may divide the fluorescence image into partial regions having constant brightness, and extract a region having a shape corresponding to a part of the detection portion 110 or the entire detection portion 110 among the partial regions as the region of the detection portion 110 .
  • the detection unit 250 generates the brightness of the region of the detection portion 110 .
  • the detection unit 250 specifies, as a value representing the brightness of the region of the detection portion 110 , a representative value (an average value, a median value, an intermediate value, a mode value, or the like) of the pixel values of the pixels included in the region of the detection portion 110 .
  • the detection unit 250 associates a measurement time and the position information with the value representing the brightness of the region of the detection portion 110 , and stores the value representing the brightness of the region of the detection portion 110 and associated with the measurement time and the position information in the measurement data storage unit 260 .
  • the detection unit 250 may individually store, in the measurement data storage unit 260 , the value of each of the plurality of detection portions 110 , the value representing the brightness of the region of the detection portion 110 and being associated with the measurement time and the position information.
  • the measurement data storage unit 260 stores the trend of the measurement result.
  • the trend of the measurement result is, for example, values representing the brightness of the region of the detection portion 110 and associated with the measurement time and the position information at a plurality of time points.
  • the measurement data storage unit 260 may store the trend of the measurement result of each of the plurality of detection portions 110 .
  • the generation unit 210 reads the trend of the measurement result from the measurement data storage unit 260 .
  • the trend of the measurement result is values representing the brightness of the region of the detection portion 110 and associated with the measurement time and the position information at a plurality of time points.
  • the generation unit 210 estimates the trend of the amount of the plant hormone that has reacted with the detection portion 110 from the trend of the measurement result.
  • the generation unit 210 specifies the detection portion 110 around which a plant infected with a disease is likely to exist, based on the trend of the amount of the plant hormone that has reacted with the detection portion 110 .
  • the generation unit 210 generates the position information of the diseased plant based on the position of the detection portion 110 specified as the detection portion 110 around which a plant infected with a disease is likely to exist.
  • the generation unit 210 calculates a change in amount of the plant hormone that has reacted with the detection portion 110 per unit period, from the amount of the plant hormone that has reacted with the detection portion 110 per unit area at least at the latest two time points in the trend of the measurement result. For example, as described above, the generation unit 210 specifies the detection portion 110 around which a plant infected with a disease is likely to exist based on the change in amount of the plant hormone that has reacted with the detection portion 110 per unit period.
  • the generation unit 210 may specify the detection portion 110 around which a plant infected with a disease is likely to exist among the plurality of detection portions 110 .
  • the generation unit 210 may generate the position information of the infected plant as in the example described in the fourth example embodiment for each detection portion 110 specified as the detection portion 110 around which a plant infected with a disease is likely to exist.
  • the generation unit 210 may output information in which a figure that indicates the position information of the infected plant and is generated for the plurality of detection portions 110 is superimposed on an image indicating a region where the plant is cultivated.
  • the mark indicating the position information of the infected plant may be the same as the mark indicating the position information of the infected plant in the fourth example embodiment.
  • the output unit 220 outputs the position information of the infected plant to the terminal device 500 .
  • the terminal device 500 includes an information receiving unit 510 , a display control unit 520 , and a display unit 530 .
  • the terminal device 500 can be implemented by, for example, a terminal device such as a smartphone, a tablet terminal, or a notebook personal computer.
  • the information receiving unit 510 receives the position information of the infected plant from the output unit 220 of the assistance device 201 .
  • the display control unit 520 controls the display unit 530 to display the received position information of the infected plant.
  • the display unit 530 displays the position information of the infected plant.
  • the display unit 530 is, for example, a display.
  • FIG. 7 is a block diagram illustrating an example of an operation of the assistance device 201 according to the fifth example embodiment of the present disclosure.
  • the measurement data acquisition unit 230 acquires the measurement data (Step S 201 ).
  • the position information acquisition unit 240 acquires the position information (Step S 202 ).
  • the detection unit 250 detects the region of the detection portion 110 from the measurement data (Step S 203 ).
  • the detection unit 250 determines the brightness of the detection portion 110 based on the pixel value of the detected region (Step S 204 ).
  • the generation unit 210 estimates the trend of the amount of the plant hormone that has reacted with the detection portion 110 based on the trend of the brightness of the region of the detection portion 110 (Step S 205 ).
  • the generation unit 210 estimates the detection portion 110 around which a plant infected with a disease is likely to exist, based on the trend of the amount of the plant hormone that has reacted with the detection portion 110 (Step S 206 ).
  • the generation unit 210 generates the position information of the plant infected with the disease (Step S 207 ).
  • the output unit 220 outputs the position information of the plant infected with the disease (Step S 208 ).
  • the detection device 100 may be attached to a support such as a pillar erected on the ground in a region where a plant is cultivated. A plurality of detection devices 100 may be attached to one pillar at different heights.
  • the detection device 100 may be attached to a support such as a pillar attached to a beam horizontally attached to a ceiling or an upper portion of a greenhouse in which a plant is cultivated.
  • the detection device 100 may be attached to a trunk, a stem, or a branch of a plant.
  • the detection device 100 may be attached to a support attached to a trunk, a stem, or a branch of a plant.
  • the detection device 100 may be a part of a structure onto which the substance is applied.
  • the detection device 100 is a tape impregnated with the substance or a tape on which the substance is applied, and may be wound around a trunk, a stem, or a branch of a plant.
  • the light source 300 and the sensor 400 may be mounted on a measurement terminal.
  • a person in charge of measurement who holds the measurement terminal may use the measurement terminal to irradiate the detection device 100 with light by using the light source 300 , and perform measurement by using the sensor 400 while the detection device 100 is irradiated with light.
  • the light source 300 and the sensor 400 may be mounted on a mobile object such as a self-propelled robot, a vehicle, or a drone.
  • the mobile object that is a self-propelled robot or a vehicle may irradiate the detection device 100 with light by using the light source 300 and perform measurement by using the sensor 400 during the irradiation of the detection device 100 with light while traveling in a passage in a region where a plant is cultivated.
  • the mobile object that is a flying object such as a drone or a helicopter may irradiate the detection device 100 with light by using the light source 300 and perform measurement by using the sensor 400 during the irradiation of the detection device 100 with light while flying over a region where a plant is cultivated.
  • the light source 300 may be installed in a place in which a plant is cultivated. While the light source 300 irradiates the detection device 100 with light, the person in charge of measurement may perform measurement by using the sensor 400 that is a digital camera, a camera mounted on a smartphone, or the like, while moving through a passage in a region where a plant is cultivated.
  • the sensor 400 that is a digital camera, a camera mounted on a smartphone, or the like
  • the position information acquisition unit 240 may receive an image that is the measurement data.
  • the position information acquisition unit 240 may derive the position of the detection device 100 (that is, the position of the detection portion 110 ) from the image.
  • a sign indicating the position may be appropriately installed at a place in which the plant is cultivated.
  • the sensor 400 captures an image in such a way that the region of the detection device 100 and the region of the sign are included in one image.
  • the position information acquisition unit 240 extracts the region of the detection device 100 (or the detection portion 110 ) and the region of the sign from the image that is the measurement data.
  • the position information acquisition unit 240 Based on the region of the sign, the position information acquisition unit 240 recognizes the position indicated by the sign by using an image recognition technology such as character recognition.
  • the position information acquisition unit 240 specifies the position of the detection device 100 (that is, the detection portion 110 ) based on a positional relationship between the region of the detection device 100 (or the detection portion 110 ) and the region of the sign, and the position indicated by the sign.
  • the position information acquisition unit 240 may calculate the positions of the detection device 100 (that is, the detection portion 110 ) and the sign in the place in which the plant is cultivated, from the image including the region of the detection device 100 (that is, the detection portion 110 ) and the region of the sign.
  • the position information acquisition unit 240 may specify the position of the detection device 100 (that is, the detection portion 110 ) based on the positional relationship between the detection device 100 (that is, the detection portion 110 ) and the sign in the place in which the plant is cultivated.
  • the position information acquisition unit 240 may specify the position of the detection device 100 (that is, the detection portion 110 ) from the image by using another method.
  • the sensor 400 may have a function of specifying latitude and longitude by using, for example, a global positioning system (GPS) or the like.
  • the sensor 400 may transmit information on the latitude and longitude specified when the detection portion 110 is measured to the position information acquisition unit 240 as the position information of the detection device 100 including the measured detection portion 110 .
  • the position information acquisition unit 240 may use the information on the latitude and longitude received from the sensor 400 as the position information of the detection device 100 including the measured detection portion 110 .
  • the sensor 400 may be one camera or a plurality of cameras fixed to a place in which a plant is cultivated (for example, in a greenhouse).
  • the detection device 101 may be used.
  • the person in charge of measurement operates the shielding portion 130 of the detection device 101 in such a way as not to shield the reference portion 120 .
  • the person in charge of measurement measures (in other words, captures an image of) the detection device 101 by using the sensor 400 in such a way that the image includes the region of the detection portion 110 and the region of the reference portion 120 that is not being shielded by the shielding portion 130 .
  • the person in charge of measurement may perform the measurement of the detection device 101 by using the sensor 400 after operating the shielding portions 130 of all the detection devices 101 in such a way as not to shield the reference portions 120 .
  • the person in charge of measurement operates the shielding portions 130 of all the detection devices 101 in such a way as to shield the reference portions 120 .
  • the detection unit 250 extracts the region of the detection portion 110 and the region of the reference portion 120 from the image that is the measurement data. Then, a value representing the brightness of the region of the detection portion 110 and a value representing the brightness of the region of the reference portion 120 are determined.
  • the value representing the brightness of the region of the reference portion 120 is, for example, a representative value (an average value, a median value, an intermediate value, a mode value, or the like) of the pixel values of the pixels included in the region of the reference portion 120 .
  • the detection unit 250 uses the same type of representative values as the value representing the brightness of the region of the detection portion 110 and the value representing the brightness of the region of the reference portion 120 .
  • the detection unit 250 stores, instead of the value representing the brightness of the region of the detection portion 110 , the value representing the brightness of the region of the detection portion 110 and the value representing the brightness of the region of the reference portion 120 in the measurement data storage unit 260 as a part of the trend of the measurement data.
  • the generation unit 210 reads the trend of the measurement data from the measurement data storage unit 260 , and calculates a difference between the value representing the brightness of the region of the detection portion 110 and the value representing the brightness of the region of the reference portion 120 in the trend of the measurement data.
  • the generation unit 210 estimates the trend of the amount of the plant hormone that has reacted with the detection portion 110 based on the amount of change in difference between the value representing the brightness of the region of the detection portion 110 and the value representing the brightness of the region of the reference portion 120 in the trend of the measurement result.
  • the generation unit 210 holds in advance a relationship between the difference between the value representing the brightness of the region of the detection portion 110 and the value representing the brightness of the region of the reference portion 120 and the amount of the plant hormone that has reacted with the detection portion 110 per unit area.
  • the generation unit 210 estimates the amount of the plant hormone that has reacted with the detection portion 110 per unit area from the calculated difference in brightness based on the held relationship.
  • the detection device 102 may be used.
  • the position information acquisition unit 240 receives an image that is the measurement data, and detects a region of the identification information portion 140 in the received image. Then, the position information acquisition unit 240 extracts the identification information indicated by the identification information portion 140 from the region of the identification information portion 140 of the image.
  • an existing method appropriate for the type of the identification information can be used.
  • the position information acquisition unit 240 specifies the position of the measured detection device 102 (that is, the position of the measured detection portion 110 ) from the extracted identification information based on a relationship between the identification information and a place in which the detection device 102 including the identification information portion 140 indicating the identification information is installed.
  • the position information acquisition unit 240 holds in advance the relationship between the identification information and the place in which the detection device 102 including the identification information portion 140 indicating the identification information is installed.
  • the position information acquisition unit 240 specifies the place in which the detection device 102 including the identification information portion 140 indicating the extracted identification information is installed, in information indicating the held relationship.
  • the sensor 400 includes, for example, a camera and an RFID reader.
  • the RFID reader of the sensor 400 reads the identification information from the identification information portion 140 that is an RFID tag.
  • the sensor 400 transmits the read identification information to the position information acquisition unit 240 .
  • the identification information portion 140 may be a module that specifies the position information (that is, the information on the latitude and longitude) by using GPS.
  • the sensor 400 acquires the specified position information from the identification information portion 140 , and transmits the acquired position information to the position information acquisition unit 240 .
  • the position information acquisition unit 240 receives, from the sensor 400 , the identification information read from the identification information portion 140 that is an RFID tag.
  • the position information acquisition unit 240 specifies the position of the measured detection device 102 (that is, the position of the measured detection portion 110 ) from the received identification information based on a relationship between the identification information and a place in which the detection device 102 including the identification information portion 140 indicating the identification information is installed. Also in this case, the position information acquisition unit 240 holds in advance the relationship between the identification information and the place in which the detection device 102 including the identification information portion 140 indicating the identification information is installed. For example, the position information acquisition unit 240 specifies the place in which the detection device 102 including the identification information portion 140 indicating the received identification information is installed, in information indicating the held relationship.
  • the detection device 100 may be stretched between a plurality of pillars.
  • the detection device 100 may be suspended from a support such as a pillar.
  • the position information acquisition unit 240 specifies, for example, the position of a portion of the detection device 100 in the image where the intensity of fluorescence is maximum.
  • the position of the portion where the intensity of fluorescence is maximum may be represented by, for example, a distance from any one end of the detection device 100 in a three-dimensional space. In this case, the position information acquisition unit 240 specifies the position of a portion of the detection device 100 in the image where the intensity of fluorescence is maximum.
  • the position information acquisition unit 240 may calculate the distance from any one end of the detection device 100 in the three-dimensional space based on a camera parameter of the camera that serves as the sensor 400 .
  • the position information acquisition unit 240 may further use the shape of the detection device 100 measured in advance to calculate the distance from any one end of the detection device 100 in the three-dimensional space.
  • the detection unit 250 determines whether the portion of the image detection device 100 in the image where the intensity of fluorescence detected by the position information acquisition unit 240 is maximum is a portion where the intensity of fluorescence detected in the past is maximum in the history of measurement results. Specifically, the detection unit 250 detects, in the history of the measurement results, a portion from which a distance to the portion where the detected intensity of fluorescence is maximum (hereinafter, also referred to as maximum intensity portion) is equal to or less than a predetermined distance and in which the detected intensity of fluorescence is maximum (hereinafter, also referred to as past maximum intensity portion).
  • the detection unit 250 stores a value representing the brightness of the maximum intensity portion in the measurement data storage unit 260 as the latest measurement result of the brightness of the detected past maximum intensity portion in the history of the measurement results. In a case where the past maximum intensity portion has not been detected, the detection unit 250 stores a value representing the brightness of the maximum intensity portion in the measurement data storage unit 260 as a new value of the brightness of the portion where the intensity of fluorescence is maximum in the history of the measurement results.
  • the light source 300 irradiates the applied substance with light via the detection device 100 that is the optical fiber.
  • the detection portion 110 In a case where the shape of the detection portion 110 represents the identification information of the detection device 102 including the detection portion 110 , the detection portion 110 also operates as the identification information portion 140 .
  • the outline of the detection portion 110 may be drawn by using a substance that emits fluorescence when irradiated with light regardless of the presence or absence of the plant hormone.
  • the shape of the detection portion 110 may be, for example, a predetermined figure, a character, or the like.
  • the position information acquisition unit 240 detects the region of the detection portion 110 from the image that is the measurement data.
  • the detection unit 250 may detect the region of the detection portion 110 .
  • the position information acquisition unit 240 recognizes the shape of the detection portion 110 detected. In this case, a shape that most matches the shape of the detection portion 110 among a plurality of predetermined shapes is estimated using any one of existing methods in which transformation such as affine transformation or perspective transformation and template matching are combined.
  • the position information acquisition unit 240 may estimate the shape of the detection portion 110 in the three-dimensional space by using any one of existing methods using the camera parameter and the like.
  • the position information acquisition unit 240 may estimate a shape that most matches the estimated shape of the detection portion 110 among a plurality of predetermined shapes.
  • the position information acquisition unit 240 holds in advance a combination of the shape of a figure and information indicating a place in which the detection device 102 including the detection portion 110 having the shape is installed, for each of a plurality of predetermined shapes. Then, the position information acquisition unit 240 determines that the place in which the detection device 102 is installed and which is combined with the shape that most matches the estimated shape of the detection portion 110 is the place in which the detection device 102 including the measured detection portion 110 is installed.
  • the position information acquisition unit 240 performs character recognition on a region including the region of the detection portion 110 by using any one of existing technologies for recognizing a character in a captured image.
  • the position information acquisition unit 240 holds in advance a combination of a character and a place in which the detection device 102 including the detection portion 110 having the shape of the character is installed, for a plurality of predetermined characters.
  • the position information acquisition unit 240 determines that the place in which the detection device 102 is installed and which is combined with the character recognized by the above-described character recognition is the place in which the detection device 102 including the detection portion 110 on which the above-described character recognition has been performed is installed.
  • the present example embodiment has the same effect as those of the fourth example embodiment.
  • the reason is the same as the reason why the effects of the fourth example embodiment are exhibited.
  • the sensor 400 may be a photodetector having sensitivity to a wavelength of fluorescence emitted by the detection portion 110 irradiated with light.
  • the measurement data is a value representing the brightness.
  • the detection unit 250 stores a value representing the brightness indicated by the measurement data acquired by the measurement data acquisition unit 230 in the measurement data storage unit 260 .
  • FIG. 8 is a block diagram illustrating an example of a configuration of a terminal device 500 according to a sixth example embodiment of the present disclosure.
  • the terminal device 500 of the present example embodiment is the same as the terminal device 500 according to the fifth example embodiment.
  • the terminal device 500 of the present example embodiment is connected to the assistance device 201 according to the fifth example embodiment.
  • the terminal device 500 includes an information receiving unit 510 , a display control unit 520 , and a display unit 530 .
  • the terminal device 500 can be implemented by, for example, a terminal device such as a smartphone, a tablet terminal, or a notebook personal computer.
  • the information receiving unit 510 receives the position information of the infected plant from the output unit 220 of the assistance device 201 .
  • the display control unit 520 controls the display unit 530 to display the received position information of the infected plant.
  • the display unit 530 displays the position information of the infected plant.
  • the display unit 530 is, for example, a display.
  • the terminal device 500 may be connected to the assistance device 200 according to the fourth example embodiment.
  • the assistance device 200 and the assistance device 201 described above can be implemented by a computer including a memory into which a program read from a storage medium is loaded and a processor that executes the program.
  • the assistance device 200 and the assistance device 201 can also be implemented by dedicated hardware.
  • the assistance device 200 and the assistance device 201 can also be implemented by a combination of the above-described computer and dedicated hardware.
  • FIG. 9 is a diagram illustrating an example of a hardware configuration of a computer 1000 that can implement the assistance devices 200 and 201 according to the above-described example embodiments.
  • the computer 1000 includes a processor 1001 , a memory 1002 , a storage device 1003 , and an input/output (I/O) interface 1004 .
  • the computer 1000 can access a storage medium 1005 .
  • the memory 1002 and the storage device 1003 are, for example, storage devices such as a random access memory (RAM) and a hard disk.
  • the storage medium 1005 is, for example, a storage device such as a RAM or a hard disk, a read only memory (ROM), or a portable storage medium.
  • the storage device 1003 may be the storage medium 1005 .
  • the processor 1001 can read and write data and programs from and in the memory 1002 and the storage device 1003 .
  • the processor 1001 can access, for example, the sensor 400 and the terminal device 500 via the I/O interface 1004 .
  • the processor 1001 can access the storage medium 1005 .
  • the storage medium 1005 stores a program for operating the computer 1000 as the assistance device according to the example embodiment of the present disclosure.
  • the processor 1001 loads, into the memory 1002 , a program that is stored in the storage medium 1005 and causes the computer 1000 to operate as the assistance device according to the example embodiment of the present disclosure. Then, the processor 1001 executes the program loaded into the memory 1002 , whereby the computer 1000 operates as the assistance device according to the example embodiment of the present disclosure.
  • the generation unit 210 , the output unit 220 , the measurement data acquisition unit 230 , the position information acquisition unit 240 , and the detection unit 250 can be implemented by, for example, the processor 1001 that executes the program loaded into the memory 1002 .
  • the measurement data storage unit 260 can be implemented by the memory 1002 included in the computer 1000 or the storage device 1003 such as a hard disk device. Some or all of the generation unit 210 , the output unit 220 , the measurement data acquisition unit 230 , the position information acquisition unit 240 , the detection unit 250 , and the measurement data storage unit 260 can be implemented by a dedicated circuit.
  • a detection device including:
  • the detection device in which the substance emits the fluorescence with an intensity that is in proportion to an amount of the reacted plant hormone.
  • the detection device according to Supplementary Note 1 or 2, further including:
  • the detection device according to any one of Supplementary Notes 1 to 3, further including:
  • An assistance device including:
  • An assistance system including:
  • a terminal device including:
  • the terminal device according to Supplementary Note 12, further including:
  • An assistance method including:
  • the present disclosure has an effect of enabling recognition of the presence of a plant with an early-stage disease.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Chemical & Material Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Immunology (AREA)
  • Theoretical Computer Science (AREA)
  • Analytical Chemistry (AREA)
  • Pathology (AREA)
  • Biochemistry (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Multimedia (AREA)
  • Chemical Kinetics & Catalysis (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Food Science & Technology (AREA)
  • Medicinal Chemistry (AREA)
  • Wood Science & Technology (AREA)
  • Botany (AREA)
  • Optics & Photonics (AREA)
  • Quality & Reliability (AREA)
  • Radiology & Medical Imaging (AREA)
  • Medical Informatics (AREA)
  • Plasma & Fusion (AREA)
  • Investigating, Analyzing Materials By Fluorescence Or Luminescence (AREA)
  • Investigating Or Analysing Materials By The Use Of Chemical Reactions (AREA)

Abstract

A detection device according to an aspect of the present disclosure includes a detection portion that contains a substance that emits fluorescence in a case of being irradiated with light when reacting with a plant hormone.

Description

This application is based upon and claims the benefit of priority from Japanese Patent Application No. 2021-122182, filed on Jul. 27, 2021, the disclosure of which is incorporated herein in its entirety by reference.
TECHNICAL FIELD
The present disclosure relates to a method for detecting plant diseases.
BACKGROUND ART
International Publication No. WO 2021/111621 discloses an example of a method for detecting plant diseases.
A pathological diagnosis system for plants described in International Publication No. WO 2021/111621 performs pathology determination based on a spot of a crop in an image captured by a camera mounted on a drone.
SUMMARY
An object of the present disclosure is to provide a device or the like enabling recognition of the presence of a plant with an early-stage disease.
A detection device according to an aspect of the present disclosure includes a detection portion that contains a substance that emits fluorescence in a case of being irradiated with light when reacting with a plant hormone.
An assistance device according to an aspect of the present disclosure includes: a generation unit that generates position information of a plant infected with a disease based on a captured image of a detection portion irradiated with light and a position of the detection portion, the detection portion containing a substance that emits fluorescence in a case of being irradiated with the light when reacting with a plant hormone; and an output unit that outputs the generated position information.
A terminal device according to an aspect of the present disclosure includes: a receiving unit that receives position information of a plant infected with a disease, the position information being generated based on a captured image of a detection portion irradiated with light and a position of the detection portion, and the detection portion containing a substance that emits fluorescence in a case of being irradiated with the light when reacting with a plant hormone; and a display unit that displays the received position information.
An assistance method according to an aspect of the present disclosure includes: generating position information of a plant infected with a disease based on a captured image of a detection portion irradiated with light and a position of the detection portion, the detection portion containing a substance that emits fluorescence in a case of being irradiated with the light when reacting with a plant hormone; and outputting the generated position information.
A non-transitory computer-readable storage medium according to an aspect of the present disclosure stores a program for causing a computer to execute: generation processing of generating position information of a plant infected with a disease based on a captured image of a detection portion irradiated with light and a position of the detection portion, the detection portion containing a substance that emits fluorescence in a case of being irradiated with the light when reacting with a plant hormone; and output processing of outputting the generated position information.
BRIEF DESCRIPTION OF THE DRAWINGS
Exemplary features and advantages of the present invention will become apparent from the following detailed description when taken with the accompanying drawings in which:
FIG. 1 is a schematic diagram schematically illustrating an example of a configuration of a detection device according to a first example embodiment of the present disclosure;
FIG. 2 is a schematic diagram schematically illustrating an example of a configuration of a detection device according to a second example embodiment of the present disclosure;
FIG. 3 is a schematic diagram schematically illustrating an example of a configuration of a detection device according to a third example embodiment of the present disclosure;
FIG. 4 is a block diagram illustrating an example of a configuration of an assistance device according to a fourth example embodiment of the present disclosure;
FIG. 5 is a flowchart illustrating an example of an operation of the assistance device according to the fourth example embodiment of the present disclosure;
FIG. 6 is a block diagram illustrating an example of a configuration of an assistance system according to a fifth example embodiment of the present disclosure;
FIG. 7 is a block diagram illustrating an example of an operation of the assistance device according to the fifth example embodiment of the present disclosure;
FIG. 8 is a block diagram illustrating an example of a configuration of a terminal device according to a sixth example embodiment of the present disclosure; and
FIG. 9 is a diagram illustrating an example of a hardware configuration of a computer that can implement the assistance devices according to the above-described example embodiments.
EXAMPLE EMBODIMENT
Hereinafter, example embodiments of the present disclosure will be described in detail with reference to the drawings.
First Example Embodiment
FIG. 1 is a schematic diagram schematically illustrating an example of a configuration of a detection device according to a first example embodiment of the present disclosure. In the example illustrated in FIG. 1 , the detection device 100 includes a detection portion 110.
The detection portion 110 shows a reaction that is in proportion to the amount of infection signal substance of a disease of a plant. Specifically, when the detection portion 110 is irradiated with light, the detection portion 110 irradiated with light emits fluorescence with an intensity that is in proportion to the amount of the infection signal substance that has reacted with the detection portion 110. In the description of the present disclosure, the infection signal substance is also referred to as a plant hormone. The detection portion 110 contains, for example, a substance whose property is changed due to the infection signal substance of a disease of a plant. The infection signal substance is, for example, a substance released from a plant infected with a disease, such as methyl salicylate or methyl jasmonate. The plant disease of the present disclosure is a disease that causes an infected plant to release the infection signal substance. In a case where a plant is infected with a disease (specifically, in a case where a plant is infected with pathogenic bacteria or the like), the plant begins to release the infection signal substance before a visible symptom (for example, a spot caused by a disease of a leaf of the plant, and the like) appears in the plant. In other words, the plant infected with the disease releases the infection signal substance without having the symptom, and then, the symptom appears in the plant (that is, a symptomatic state). Therefore, the present example embodiment has an effect that a plant infected with a disease can be detected before a visible symptom appears in the plant due to the disease.
The substance contained in the detection portion 110 (that is, a substance that emits fluorescence with an intensity that is in proportion to the amount of the reacted infection signal substance such as methyl salicylate or methyl jasmonate) is, for example, at least one of the following substances. Among these substances, a rare earth and a zinc compound emit fluorescence in a case of being irradiated with ultraviolet light having a wavelength of 200 to 400 nanometers (nm). A hydrazine derivative emits fluorescence in a case of being irradiated with ultraviolet light or visible light having a wavelength of 250 to 500 nm. In other words, in a case where, among the substances exemplified below, the rare earth and the zinc compound are used in the detection portion 110, the light with which the detection portion 110 is irradiated is light including ultraviolet light having a wavelength of 200 to 400 nm. In a case where, among the substances exemplified below, the hydrazine derivative is used in the detection portion 110, the light with which the detection portion 110 is irradiated is light including at least one of ultraviolet light or visible light having a wavelength of 250 to 500 nm.
First Substance
A first substance is a boron-oxygen compound having a structure represented by the following General Formula (1) as a receptor that selectively recognizes methyl salicylate.
Figure US12298245-20250513-C00001
The boron-oxygen compound may be a diboronic acid derivative represented by the following General Formula (2). (In the following formula, R represents a methyl group, an ethyl group, a propyl group, a butyl group, or a phenyl group, and each R may be the same or different.)
Figure US12298245-20250513-C00002
The boron-oxygen compound may be a phenylboronic acid derivative represented by the following General Formula (3). (In the following formula, X represents a hydrogen molecule, an alkyl group having 1 to 4 carbon atoms, an alkoxy group, a mercapto group, a mercapto-substituted alkyl group, or a mercapto-substituted alkoxy group, n represents an integer of 1 to 5, and when n is equal to or more than 2, each X may be the same or different.)
Figure US12298245-20250513-C00003
The boron-oxygen compound may be a boroxine derivative represented by the following General Formula (4). (In the following formula, Z represents an alkyl group having 1 to 4 carbon atoms or a substituted or unsubstituted phenyl group, and each Z may be the same or different.)
Figure US12298245-20250513-C00004
The first substance emits fluorescence as methyl salicylate reacts with the boron-oxygen compound to form a boron complex.
Second Substance
A second substance is a substance in which the receptor that selectively recognizes methyl salicylate is a rare earth compound.
The above-described rare earth compound may be a compound of any of an acetate, a chloride, an oxalate, a nitrate, a propionate, an isobutyrate, and a pivalate of scandium (Sc), yttrium (Y), lanthanum (La), cerium (Ce), praseodymium (Pr), neodymium (Nd), promethium (Pm), samarium (Sm), europium (Eu), gadolinium (Gd), terbium (Tb), dysprosium (Dy), holmium (Ho), erbium (Er), thulium (Tm), ytterbium (Yb), and lutetium (Lu).
The above-described rare earth compound may be a compound of any of an acetate, a chloride, an oxalate, a nitrate, a propionate, an isobutyrate, or a pivalate of samarium (Sm), europium (Eu), gadolinium (Gd), terbium (Tb), or dysprosium (Dy).
The above-described rare earth compound may be a compound in which a salt of a rare earth element forms a complex with a phosphine oxide derivative.
The above-described rare earth compound emits fluorescence as methyl salicylate reacts with the rare earth compound to form a complex of rare earth.
Third Substance
A third substance is a substance in which the receptor that selectively recognizes methyl salicylate is a zinc compound.
The above-described zinc compound is a compound of any of zinc(II) acetate, zinc(II) formate dihydrate, zinc(II) butyrate, zinc oxalate dihydrate, zinc(II) hexanoate, zinc(II) propionate, zinc(II) tartrate dihydrate, zinc(II) benzoate, zinc(II) octanoate, zinc(II) oleate, zinc(II) nitrate, and zinc(II) chloride.
The above-described zinc compound emits fluorescence as methyl salicylate reacts with the zinc compound to form a complex.
Fourth Substance
A fourth substance is a substance in which the receptor that selectively recognizes methyl jasmonate is a hydrazine derivative.
The above-described hydrazine derivative is a compound of any of 4-hydrazino-7-nitro-2,1,3-benzoxadiazole hydrazine, 5-dimethylaminonaphthalene-1-sulfonyl hydrazine, and 4-(N,N-dimethylaminosulfonyl)-7-hydrazino-2,1,3-benzoxadiazole.
The above-described hydrazine derivative emits fluorescence as methyl jasmonate reacts with the hydrazine derivative to form a hydrazone derivative.
These substances emit fluorescence with an intensity that is in proportion to the amount of methyl salicylate that has reacted with these substances per unit weight in a case of being irradiated with light. The detection portion 110 contains at least one of these substances.
Effects
The present example embodiment has an effect of enabling recognition of the presence of a plant with an early-stage disease. This is because, in a case of being irradiated with light, the detection portion 110 emits fluorescence with an intensity based on the amount of the infection signal substance of a disease of a plant that has reacted with the detection portion 110 per unit area.
Specific Example of First Example Embodiment
The detection device 100 can be implemented by, for example, a base such as a plate to which the detection portion 110 such as a filter paper impregnated with the substance is attached. The material of the base may be a combination of one or more of wood, metal, resin, mineral, and the like. The material of the base is not particularly limited. The shape of the base is not particularly limited. The base may be formed as, for example, a plate, a column, a thread, a string, a cable, a tape, a sheet, or the like. The detection portion 110 may be, for example, an object impregnated with the substance. In this case, the detection portion 110 may be, for example, a filter paper, a cloth, a resin sheet that can be impregnated with the substance, or the like. The detection portion 110 may be the substance applied onto the base or a mixture of the substance and another substance applied onto the base. The shape of the detection portion 110 is not particularly limited. The shape of the detection portion 110 may be a predetermined figure such as a rectangle, a polygon, or a circle. The shape of the detection portion 110 may be a contour line of a predetermined figure having a finite thickness. The shape of the detection portion 110 may be a mark, a character, a character string, or the like. The shape of the detection portion 110 may be a graphic such as a barcode or a quick response (QR) code (registered trademark).
Hereinafter, a more specific example of the detection device 100 will be described.
The detection device 100 may be a resin substrate, a wood substrate, a metal substrate, or the like to which a sheet such as a filter paper impregnated with the substance is attached. The detection device 100 may be a substrate onto which a substance containing the substance is applied. The detection device 100 may be a part of a structure such as a garden stone or a wall to which a sheet such as a filter paper impregnated with the substance is attached. The detection device 100 may be a part of a structure such as a garden stone or a wall onto which a substance including the substance is applied. The detection device 100 may be a tape onto which the substance is applied. The material of the tape is, for example, paper, resin, a combination of paper and resin, or another material that can form the tape. The detection device 100 may be a tape impregnated with the substance. The detection device 100 may be a thread, a string, a cable, an optical fiber, or the like onto which the substance is applied. The optical fiber is an optical fiber configured in such a way that a part of light passing through the inside leaks from the surface. The detection device 100 may be a thread or a string impregnated with the substance. The detection device 100 may be a cable or an optical fiber coated with a substance impregnated with the substance. In a case where the detection device 100 has a portion impregnated with the substance, the portion impregnated with the substance functions as the detection portion 110. In a case where the detection device 100 has a portion onto which the substance is applied, the portion onto which the substance is applied functions as the detection portion 110.
The outline of the detection portion 110 may be drawn by using a fluorescent substance that emits fluorescence when irradiated with light.
Second Example Embodiment
FIG. 2 is a schematic diagram schematically illustrating an example of a configuration of a detection device according to a second example embodiment of the present disclosure. In the example illustrated in FIG. 2 , a detection device 101 includes a detection portion 110, a reference portion 120, and a shielding portion 130. The detection portion 110 and the reference portion 120 of the present example embodiment are the same as the detection portion 110 of the first example embodiment.
The shielding portion 130 is a detachable component that shields the reference portion 120 from the outside air containing the infection signal substance. Specifically, the shielding portion 130 covers the reference portion 120 in such a way that the reference portion 120 is shielded from the outside air except when the shielding portion 130 is removed from the reference portion 120. The shielding portion 130 can suppress a reaction between the substance in the reference portion 120 and the infection signal substance contained in the outside air by covering the reference portion 120. In the description of the present example embodiment, the outside air is air other than air between the shielding portion 130 and the reference portion 120.
Effects
The present example embodiment has the same effects as those of the first example embodiment. The reason is the same as the reason why the effects of the first example embodiment are exhibited.
Specific Example of Second Example Embodiment
The detection device 101 according to the second example embodiment of the present disclosure has a configuration similar to that of the detection device 100 according to the first example embodiment. However, in the portion impregnated with the substance, a portion that is not shielded by the shielding portion 130 may function as the detection portion 110, and a portion that is shielded by the shielding portion 130 may function as the reference portion 120. In the portion onto which the substance is applied, a portion that is not shielded by the shielding portion 130 may function as the detection portion 110, and a portion that is shielded by the shielding portion 130 may function as the reference portion 120. The detection portion 110 and the reference portion 120 may be formed to be separated from each other. The detection portion 110 and the reference portion 120 may be integrally formed.
As described above, the shielding portion 130 is a component that covers the reference portion 120 in such a way that the reference portion 120 is shielded from the outside air, and is detachable from the reference portion 120. The material and shape of the shielding portion 130 are not particularly limited.
The outline of the detection portion 110 may be drawn by using a fluorescent substance that emits fluorescence when irradiated with light. The outline of the reference portion 120 may be drawn by using a fluorescent substance that emits fluorescence when irradiated with light.
Third Example Embodiment
FIG. 3 is a schematic diagram schematically illustrating an example of a configuration of a detection device 102 according to a third example embodiment of the present disclosure. The detection device 102 includes a detection portion 110 and an identification information portion 140. The detection device 102 may further include a reference portion 120 and a shielding portion 130. The detection portion 110 is the same as the detection portions 110 of the first example embodiment and the second example embodiment. The reference portion 120 and the shielding portion 130 are the same as the reference portion 120 and the shielding portion 130 of the second example embodiment.
The identification information portion 140 holds identification information assigned to the detection device 102 in such a way that the identification information can be read from the outside of the detection device 102. The identification information is information that is assigned to the detection device 102 and can individually specify the detection device 102. The identification information may be, for example, a number such as a serial number. The identification information may be represented by a character string. The identification information may be represented by data. The identification information may be represented by a barcode, a QR code, or the like.
Effects
The present example embodiment has the same effects as those of the first example embodiment. The reason is the same as the reason why the effects of the first example embodiment are exhibited.
Specific Example of Third Example Embodiment
The detection portion 110 of the present example embodiment can be implemented by the specific example of the detection portion 110 described as the specific example of the first example embodiment. The reference portion 120 and the shielding portion 130 of the present example embodiment can be implemented by the reference portion 120 and the shielding portion 130 described as the specific example of the second example embodiment.
The identification information portion 140 may be a portion of the surface of the detection device 102 on which the identification information is drawn. In this case, the identification information is, for example, a number or another character string, a figure, a barcode, a QR code, or the like. In this case, the identification information may be drawn by using a fluorescent substance. The identification information may be drawn by using the substance. In a case where the identification information is drawn by using the substance, the identification information portion 140 may also function as the detection portion 110. The identification information portion 140 may be a radio frequency identifier (RFID) tag holding the identification information.
Fourth Example Embodiment
Hereinafter, a fourth example embodiment of the present disclosure will be described in detail with reference to the drawings.
Configuration
FIG. 4 is a block diagram illustrating an example of a configuration of an assistance device 200 according to the fourth example embodiment of the present disclosure. In the example illustrated in FIG. 4 , the assistance device 200 includes a generation unit 210 and an output unit 220.
The generation unit 210 generates position information of a plant infected with a disease based on a captured image of the detection portion 110 irradiated with light and a position of the detection portion 110, the detection portion 110 containing a substance that emits fluorescence in a case of being irradiated with the light when reacting with a plant hormone. The detection portion is the detection portion 110 according to any one of the first to third example embodiments described above.
The output unit 220 outputs the generated position information.
For example, the generation unit 210 may estimate information on the amount of the plant hormone at the position of the detection portion 110 from the brightness of the detection portion 110 in the captured image based on a relationship between the brightness of the detection portion 110 irradiated with ultraviolet light and the amount of the plant hormone that has reacted with the detection portion 110. The amount of the plant hormone that has reacted with the detection portion 110 may be, for example, the amount of the plant hormone that has reacted with the detection portion 110 per unit area. For example, the position of the detection device 100 including the detection portion 110 is regarded as the position of the detection portion 110. The amounts of the plant hormone at a plurality of time points are expressed as the trend of the amount of the plant hormone.
In the trend of the amount of the plant hormone that has reacted with the detection portion 110, in a case where a change in amount of the plant hormone exceeds a predetermined value, there is a possibility that a plant infected with a disease exists around the detection portion 110.
For example, the generation unit 210 specifies the detection portion 110 (that is, the detection device 100) around which a plant infected with a disease is likely to exist based on the trend of the amount of the plant hormone that has reacted with the detection portion 110. The generation unit 210 generates the position information of the plant infected with the disease based on the specified position of the detection portion 110.
For example, in a case where the change in amount of the plant hormone exceeds the predetermined value in the trend of the amount of the plant hormone that has reacted with the detection portion 110, the generation unit 210 may specify the detection portion 110 as the detection portion 110 around which a plant infected with a disease is likely to exist. For example, in a case where an increase of the change in amount of the plant hormone exceeds a predetermined value in the trend of the amount of the plant hormone that has reacted with the detection portion 110, the generation unit 210 may specify the detection portion 110 as the detection portion 110 around which a plant infected with a disease is likely to exist. In the following description, the detection portion 110 around which a plant infected with a disease is likely to exist is also referred to as a detection receptor.
The generation unit 210 generates a range based on the position of the detection receptor as information indicating a region where a plant infected with a disease is likely to exist, in other words, the position information of the plant infected with the disease.
The generation unit 210 may generate information indicating the position of the detection receptor (for example, the position of the detection device 100 including the detection receptor) as the position information of the plant infected with the disease. Specifically, the generation unit 210 generates, as the position information of the plant infected with the disease, information of a screen in which a mark (for example, a combination of one or more of a figure, a symbol, a character, a character string, and the like) indicating the position of the detection receptor is superimposed on an image indicating a place in which the plant is cultivated.
For example, the generation unit 210 may specify a region based on a distance from the position of the detection receptor (for example, a place in which the detection device 100 including the detection receptor is installed) as the region where a plant infected with a disease is likely to exist.
In this case, the region based on the distance from the position of the detection receptor may be, for example, a region within a predetermined distance from the position of the detection receptor. The region based on the distance from the position of the detection receptor may be, for example, a region that is at a first predetermined distance or more from the position of the detection receptor and is at a second predetermined distance or less from the position of the detection receptor, the second predetermined distance being larger than the first predetermined distance.
The generation unit 210 may estimate a distance from the detection portion 110 to the plant infected with the disease based on the amount of change in amount of the plant hormone in the trend of the amount of the plant hormone that has reacted with the detection portion 110. In this case, the generation unit 210 holds in advance a relationship between the amount of change in amount of the plant hormone and the distance from the detection portion to the plant infected with the disease, and estimates the distance from the detection portion 110 to the plant infected with the disease based on the relationship.
For example, the generation unit 210 may estimate the distance from the detection portion 110 to the plant infected with the disease based on an increment of the change in amount of the plant hormone in the trend of the amount of the plant hormone that has reacted with the detection portion 110. In this case, the generation unit 210 holds in advance a relationship between the increment of the change in amount of the plant hormone and the distance from the detection portion to the plant infected with the disease, and estimates the distance from the detection portion 110 to the plant infected with the disease based on the relationship.
In this case, the region based on the distance from the position of the detection receptor may be, for example, a region within the estimated distance from the position of the detection receptor. The region based on the distance from the position of the detection receptor may be, for example, a region that is at a third predetermined distance or more from the position of the detection receptor and is at a fourth predetermined distance or less from the position of the detection receptor. However, the third predetermined distance is smaller than the estimated distance, and the fourth predetermined distance is larger than the estimated distance.
For example, the generation unit 210 may specify, among a plurality of divided sections of the place in which the plant is cultivated, a section including a region within a predetermined distance from a place in which the detection device 100 including the detection receptor is installed, as the region where a plant infected with a disease is likely to exist.
In the above case, the generation unit 210 generates, as the position information of the plant infected with the disease, information indicating a region where the plant infected with the disease is likely to exist. Specifically, the generation unit 210 generates information of a screen in which a mark indicating the region where the plant infected with the disease is likely to exist is superimposed on an image indicating the place in which the plant is cultivated. The mark indicating the region where the plant infected with the disease is likely to exist may be a line surrounding the region where the plant infected with the disease is likely to exist. The mark indicating the region where the plant infected with the disease is likely to exist may be a predetermined pattern drawn on the region where the plant infected with the disease is likely to exist. The mark indicating the region where the plant infected with the disease is likely to exist may be changing the color of the region where the plant infected with the disease is likely to exist. The mark indicating the region where the plant infected with the disease is likely to exist is not limited to these examples.
The position information of the infected plant is not limited to the above example as long as it is information indicating a position, a figure, a range, and the like, for example. The position information of the infected plant may be text data or the like.
In the present example embodiment and another example embodiment of the present disclosure, a place (and a region) where a plant is cultivated may be the inside of a greenhouse. The place in which the plant is cultivated may be outdoors.
Operation
FIG. 5 is a flowchart illustrating an example of an operation of the assistance device 200 according to the fourth example embodiment of the present disclosure. In the example illustrated in FIG. 5 , the generation unit 210 generates the position information of a plant infected with a disease based on a captured image of the detection portion 110 and the position of the detection portion 110 (Step S101). Next, the output unit 220 outputs the position information of the plant infected with the disease (Step S102).
Effects
The present example embodiment has an effect of enabling recognition of the presence of a plant with an early-stage disease. This is because the generation unit 210 generates the position information of the plant infected with the disease based on the captured image of the detection portion 110 and the position of the detection portion 110. In a case of being irradiated with light, the detection portion 110 emits fluorescence with an intensity based on the amount of the infection signal substance of the disease of the plant that has reacted with the detection portion 110 per unit area.
Fifth Example Embodiment
FIG. 6 is a block diagram illustrating an example of a configuration of an assistance system according to a fifth example embodiment of the present disclosure. In the example illustrated in FIG. 6 , an assistance system 1 includes a plurality of detection devices 100, an assistance device 201, a light source 300, a sensor 400, and a terminal device 500. The assistance device 201 is communicably connected to each of the sensor 400 and the terminal device 500.
The detection device 100 is the detection device 100 according to the first example embodiment.
The light source 300 emits light. The light source 300 is disposed in such a way as to irradiate a detection portion 110 of the detection device 100 with light.
The sensor 400 is, for example, a sensor that measures the brightness of the detection portion 110 of the detection device 100 irradiated with light from the light source 300, obtained by imaging. Specifically, the sensor 400 is, for example, an imaging device that captures an image of the detection device 100 irradiated with light from the light source 300 (specifically, an image of at least a part of the detection portion 110 irradiated with light). The imaging device is, for example, a digital camera, a camera of a smartphone, or the like. The sensor 400 has a function of transmitting a measurement result (for example, an image) to the assistance device 201. The image is a moving image or one or more still images. For example, a worker performs measurement on the detection device 100 by using the sensor 400.
The sensor 400 may be disposed in such a way as to capture an image of the detection device 100 including at least a part of a region of the detection portion 110. The sensor 400 may be disposed in such a way as to capture an image including only the region of the detection portion 110. Note that parameters that affect a relationship between the brightness and color of the detection portion 110 and the brightness and color of the detection portion 110 in the image, such as the aperture, sensitivity, gain, and color temperature of the camera of the sensor 400, are fixed.
<Assistance Device 201>
The assistance device 201 includes a measurement data acquisition unit 230, a position information acquisition unit 240, a detection unit 250, a measurement data storage unit 260, a generation unit 210, and an output unit 220. The generation unit 210 and the output unit 220 of the present example embodiment are the same as the generation unit 210 and the output unit 220 of the fourth example embodiment except for differences described below.
<Measurement Data Acquisition Unit 230>
The measurement data acquisition unit 230 acquires measurement data from the sensor 400. The measurement data is, for example, an image of the detection portion 110 of the detection device 100 obtained by the sensor 400. The measurement data acquisition unit 230 transmits the acquired measurement data (for example, an image) to the detection unit 250.
The measurement data acquisition unit 230 of the present example embodiment may acquire the measurement data of each of the plurality of detection portions 110.
<Position Information Acquisition Unit 240>
The position information acquisition unit 240 acquires the position of the detection portion 110. The position information acquisition unit 240 may acquire the position of the detection device 100 as the position of the detection portion 110. In the present example embodiment, the position information acquisition unit 240 acquires the position information of each of the plurality of detection portions 110.
In the present example embodiment, the position information acquisition unit 240 may acquire, via an input device such as a keyboard or a touch panel, information indicating the position of the detection portion 110, the information being input by an operator of the assistance system 1 using the input device. The information indicating the position of the detection portion 110 may be, for example, a number assigned to the detection portion 110. The information indicating the position of the detection portion 110 may be, for example, a number assigned to the detection device 100 including the detection portion 110. In these cases, a combination of a number that is information indicating a position and position information to which the number is assigned is assigned in advance to the position information acquisition unit 240. Then, the position information acquisition unit 240 specifies, as the position information of the detection portion 110, position information to which a number indicating the acquired position is assigned. The position information may be identification information of a section of a plant cultivation field. The position information may be coordinates in a coordinate system defined in the plant cultivation field. The input information indicating the position of the detection portion 110 may be the position information of the detection portion 110 such as the identification information of the section or the coordinates.
The position information acquisition unit 240 transmits the position information of the detection portion 110 to the detection unit 250. For example, the position information acquisition unit 240 may transmit, as the position information of the detection portion 110, the position information of the detection device 100 to the detection unit 250.
<Detection Unit 250>
The detection unit 250 receives the measurement data (for example, an image) from the measurement data acquisition unit 230. Furthermore, the detection unit 250 receives the position information of the detection portion 110 from the position information acquisition unit 240.
The detection unit 250 detects a region of the detection portion 110 in the received image. In a case where the sensor 400 that is a camera is disposed in such a way as to capture an image including only the region of the detection portion 110, the detection unit 250 detects the entire image as the region of the detection portion 110. Otherwise, for example, the detection unit 250 may calculate the intensity that is in proportion to the intensity of a wavelength (for example, 545 to 546 nm) of fluorescence, from a pixel value of each pixel of the image that is the measurement data. An image in which a pixel shows the intensity of the wavelength of fluorescence is referred to as a fluorescence image. The detection unit 250 extracts the region of the detection portion 110 in the fluorescence image based on the shape of the detection portion 110. In this case, for example, the detection unit 250 may divide the fluorescence image into partial regions having constant brightness, and extract a region having a shape corresponding to a part of the detection portion 110 or the entire detection portion 110 among the partial regions as the region of the detection portion 110.
The detection unit 250 generates the brightness of the region of the detection portion 110. For example, the detection unit 250 specifies, as a value representing the brightness of the region of the detection portion 110, a representative value (an average value, a median value, an intermediate value, a mode value, or the like) of the pixel values of the pixels included in the region of the detection portion 110. Then, the detection unit 250 associates a measurement time and the position information with the value representing the brightness of the region of the detection portion 110, and stores the value representing the brightness of the region of the detection portion 110 and associated with the measurement time and the position information in the measurement data storage unit 260.
In the present example embodiment, the detection unit 250 may individually store, in the measurement data storage unit 260, the value of each of the plurality of detection portions 110, the value representing the brightness of the region of the detection portion 110 and being associated with the measurement time and the position information.
<Measurement Data Storage Unit 260>
The measurement data storage unit 260 stores the trend of the measurement result. The trend of the measurement result is, for example, values representing the brightness of the region of the detection portion 110 and associated with the measurement time and the position information at a plurality of time points. In the present example embodiment, the measurement data storage unit 260 may store the trend of the measurement result of each of the plurality of detection portions 110.
<Generation Unit 210>
The generation unit 210 reads the trend of the measurement result from the measurement data storage unit 260. As described above, the trend of the measurement result is values representing the brightness of the region of the detection portion 110 and associated with the measurement time and the position information at a plurality of time points.
The generation unit 210 estimates the trend of the amount of the plant hormone that has reacted with the detection portion 110 from the trend of the measurement result. The generation unit 210 specifies the detection portion 110 around which a plant infected with a disease is likely to exist, based on the trend of the amount of the plant hormone that has reacted with the detection portion 110. The generation unit 210 generates the position information of the diseased plant based on the position of the detection portion 110 specified as the detection portion 110 around which a plant infected with a disease is likely to exist.
Specifically, for example, the generation unit 210 calculates a change in amount of the plant hormone that has reacted with the detection portion 110 per unit period, from the amount of the plant hormone that has reacted with the detection portion 110 per unit area at least at the latest two time points in the trend of the measurement result. For example, as described above, the generation unit 210 specifies the detection portion 110 around which a plant infected with a disease is likely to exist based on the change in amount of the plant hormone that has reacted with the detection portion 110 per unit period.
In the present example embodiment, the generation unit 210 may specify the detection portion 110 around which a plant infected with a disease is likely to exist among the plurality of detection portions 110. The generation unit 210 may generate the position information of the infected plant as in the example described in the fourth example embodiment for each detection portion 110 specified as the detection portion 110 around which a plant infected with a disease is likely to exist. In the present example embodiment, the generation unit 210 may output information in which a figure that indicates the position information of the infected plant and is generated for the plurality of detection portions 110 is superimposed on an image indicating a region where the plant is cultivated. The mark indicating the position information of the infected plant may be the same as the mark indicating the position information of the infected plant in the fourth example embodiment.
<Output Unit 220>
The output unit 220 outputs the position information of the infected plant to the terminal device 500.
<Terminal Device 500>
The terminal device 500 includes an information receiving unit 510, a display control unit 520, and a display unit 530. The terminal device 500 can be implemented by, for example, a terminal device such as a smartphone, a tablet terminal, or a notebook personal computer.
The information receiving unit 510 receives the position information of the infected plant from the output unit 220 of the assistance device 201. The display control unit 520 controls the display unit 530 to display the received position information of the infected plant. The display unit 530 displays the position information of the infected plant. The display unit 530 is, for example, a display.
Operation
FIG. 7 is a block diagram illustrating an example of an operation of the assistance device 201 according to the fifth example embodiment of the present disclosure. In the example illustrated in FIG. 7 , the measurement data acquisition unit 230 acquires the measurement data (Step S201). The position information acquisition unit 240 acquires the position information (Step S202). Then, the detection unit 250 detects the region of the detection portion 110 from the measurement data (Step S203). The detection unit 250 determines the brightness of the detection portion 110 based on the pixel value of the detected region (Step S204).
The generation unit 210 estimates the trend of the amount of the plant hormone that has reacted with the detection portion 110 based on the trend of the brightness of the region of the detection portion 110 (Step S205). The generation unit 210 estimates the detection portion 110 around which a plant infected with a disease is likely to exist, based on the trend of the amount of the plant hormone that has reacted with the detection portion 110 (Step S206). The generation unit 210 generates the position information of the plant infected with the disease (Step S207). The output unit 220 outputs the position information of the plant infected with the disease (Step S208).
Another Specific Example of Fifth Example Embodiment
The detection device 100 may be attached to a support such as a pillar erected on the ground in a region where a plant is cultivated. A plurality of detection devices 100 may be attached to one pillar at different heights. The detection device 100 may be attached to a support such as a pillar attached to a beam horizontally attached to a ceiling or an upper portion of a greenhouse in which a plant is cultivated. The detection device 100 may be attached to a trunk, a stem, or a branch of a plant. The detection device 100 may be attached to a support attached to a trunk, a stem, or a branch of a plant. The detection device 100 may be a part of a structure onto which the substance is applied. The detection device 100 is a tape impregnated with the substance or a tape on which the substance is applied, and may be wound around a trunk, a stem, or a branch of a plant.
The light source 300 and the sensor 400 may be mounted on a measurement terminal. For example, a person in charge of measurement who holds the measurement terminal may use the measurement terminal to irradiate the detection device 100 with light by using the light source 300, and perform measurement by using the sensor 400 while the detection device 100 is irradiated with light.
The light source 300 and the sensor 400 may be mounted on a mobile object such as a self-propelled robot, a vehicle, or a drone. The mobile object that is a self-propelled robot or a vehicle may irradiate the detection device 100 with light by using the light source 300 and perform measurement by using the sensor 400 during the irradiation of the detection device 100 with light while traveling in a passage in a region where a plant is cultivated. The mobile object that is a flying object such as a drone or a helicopter may irradiate the detection device 100 with light by using the light source 300 and perform measurement by using the sensor 400 during the irradiation of the detection device 100 with light while flying over a region where a plant is cultivated.
The light source 300 may be installed in a place in which a plant is cultivated. While the light source 300 irradiates the detection device 100 with light, the person in charge of measurement may perform measurement by using the sensor 400 that is a digital camera, a camera mounted on a smartphone, or the like, while moving through a passage in a region where a plant is cultivated.
In a case where the sensor 400 is a camera, the position information acquisition unit 240 may receive an image that is the measurement data. The position information acquisition unit 240 may derive the position of the detection device 100 (that is, the position of the detection portion 110) from the image. In this case, a sign indicating the position may be appropriately installed at a place in which the plant is cultivated. In this case, the sensor 400 captures an image in such a way that the region of the detection device 100 and the region of the sign are included in one image. The position information acquisition unit 240 extracts the region of the detection device 100 (or the detection portion 110) and the region of the sign from the image that is the measurement data. Based on the region of the sign, the position information acquisition unit 240 recognizes the position indicated by the sign by using an image recognition technology such as character recognition. The position information acquisition unit 240 specifies the position of the detection device 100 (that is, the detection portion 110) based on a positional relationship between the region of the detection device 100 (or the detection portion 110) and the region of the sign, and the position indicated by the sign. Specifically, the position information acquisition unit 240 may calculate the positions of the detection device 100 (that is, the detection portion 110) and the sign in the place in which the plant is cultivated, from the image including the region of the detection device 100 (that is, the detection portion 110) and the region of the sign. Then, the position information acquisition unit 240 may specify the position of the detection device 100 (that is, the detection portion 110) based on the positional relationship between the detection device 100 (that is, the detection portion 110) and the sign in the place in which the plant is cultivated. The position information acquisition unit 240 may specify the position of the detection device 100 (that is, the detection portion 110) from the image by using another method.
Note that the sensor 400 may have a function of specifying latitude and longitude by using, for example, a global positioning system (GPS) or the like. In this case, the sensor 400 may transmit information on the latitude and longitude specified when the detection portion 110 is measured to the position information acquisition unit 240 as the position information of the detection device 100 including the measured detection portion 110. The position information acquisition unit 240 may use the information on the latitude and longitude received from the sensor 400 as the position information of the detection device 100 including the measured detection portion 110.
The sensor 400 may be one camera or a plurality of cameras fixed to a place in which a plant is cultivated (for example, in a greenhouse).
Instead of the detection device 100, the detection device 101 according to the second example embodiment may be used.
In this case, the person in charge of measurement operates the shielding portion 130 of the detection device 101 in such a way as not to shield the reference portion 120. Then, the person in charge of measurement measures (in other words, captures an image of) the detection device 101 by using the sensor 400 in such a way that the image includes the region of the detection portion 110 and the region of the reference portion 120 that is not being shielded by the shielding portion 130. The person in charge of measurement may perform the measurement of the detection device 101 by using the sensor 400 after operating the shielding portions 130 of all the detection devices 101 in such a way as not to shield the reference portions 120. Once the measurement of the detection device 101 is completed, the person in charge of measurement operates the shielding portions 130 of all the detection devices 101 in such a way as to shield the reference portions 120.
The detection unit 250 extracts the region of the detection portion 110 and the region of the reference portion 120 from the image that is the measurement data. Then, a value representing the brightness of the region of the detection portion 110 and a value representing the brightness of the region of the reference portion 120 are determined. The value representing the brightness of the region of the reference portion 120 is, for example, a representative value (an average value, a median value, an intermediate value, a mode value, or the like) of the pixel values of the pixels included in the region of the reference portion 120. The detection unit 250 uses the same type of representative values as the value representing the brightness of the region of the detection portion 110 and the value representing the brightness of the region of the reference portion 120. The detection unit 250 stores, instead of the value representing the brightness of the region of the detection portion 110, the value representing the brightness of the region of the detection portion 110 and the value representing the brightness of the region of the reference portion 120 in the measurement data storage unit 260 as a part of the trend of the measurement data.
The generation unit 210 reads the trend of the measurement data from the measurement data storage unit 260, and calculates a difference between the value representing the brightness of the region of the detection portion 110 and the value representing the brightness of the region of the reference portion 120 in the trend of the measurement data. The generation unit 210 estimates the trend of the amount of the plant hormone that has reacted with the detection portion 110 based on the amount of change in difference between the value representing the brightness of the region of the detection portion 110 and the value representing the brightness of the region of the reference portion 120 in the trend of the measurement result.
Specifically, for example, the generation unit 210 holds in advance a relationship between the difference between the value representing the brightness of the region of the detection portion 110 and the value representing the brightness of the region of the reference portion 120 and the amount of the plant hormone that has reacted with the detection portion 110 per unit area. The generation unit 210 estimates the amount of the plant hormone that has reacted with the detection portion 110 per unit area from the calculated difference in brightness based on the held relationship.
Instead of the detection device 100, the detection device 102 according to the third example embodiment may be used.
For example, in a case where the sensor 400 is a camera, and the camera can measure the identification information portion 140, the position information acquisition unit 240 receives an image that is the measurement data, and detects a region of the identification information portion 140 in the received image. Then, the position information acquisition unit 240 extracts the identification information indicated by the identification information portion 140 from the region of the identification information portion 140 of the image. As a method of extracting the identification information indicated by the identification information portion 140, an existing method appropriate for the type of the identification information can be used. The position information acquisition unit 240 specifies the position of the measured detection device 102 (that is, the position of the measured detection portion 110) from the extracted identification information based on a relationship between the identification information and a place in which the detection device 102 including the identification information portion 140 indicating the identification information is installed. The position information acquisition unit 240 holds in advance the relationship between the identification information and the place in which the detection device 102 including the identification information portion 140 indicating the identification information is installed. For example, the position information acquisition unit 240 specifies the place in which the detection device 102 including the identification information portion 140 indicating the extracted identification information is installed, in information indicating the held relationship.
In a case where the identification information portion 140 is an RFID tag, the sensor 400 includes, for example, a camera and an RFID reader. The RFID reader of the sensor 400 reads the identification information from the identification information portion 140 that is an RFID tag. In this case, the sensor 400 transmits the read identification information to the position information acquisition unit 240.
The identification information portion 140 may be a module that specifies the position information (that is, the information on the latitude and longitude) by using GPS. In this case, the sensor 400 acquires the specified position information from the identification information portion 140, and transmits the acquired position information to the position information acquisition unit 240.
The position information acquisition unit 240 receives, from the sensor 400, the identification information read from the identification information portion 140 that is an RFID tag. The position information acquisition unit 240 specifies the position of the measured detection device 102 (that is, the position of the measured detection portion 110) from the received identification information based on a relationship between the identification information and a place in which the detection device 102 including the identification information portion 140 indicating the identification information is installed. Also in this case, the position information acquisition unit 240 holds in advance the relationship between the identification information and the place in which the detection device 102 including the identification information portion 140 indicating the identification information is installed. For example, the position information acquisition unit 240 specifies the place in which the detection device 102 including the identification information portion 140 indicating the received identification information is installed, in information indicating the held relationship.
In a case where the detection device 100 is a thread, a string, a tape, a cable, an optical fiber that leaks light, or the like, the detection device 100 may be stretched between a plurality of pillars. The detection device 100 may be suspended from a support such as a pillar. The position information acquisition unit 240 specifies, for example, the position of a portion of the detection device 100 in the image where the intensity of fluorescence is maximum. The position of the portion where the intensity of fluorescence is maximum may be represented by, for example, a distance from any one end of the detection device 100 in a three-dimensional space. In this case, the position information acquisition unit 240 specifies the position of a portion of the detection device 100 in the image where the intensity of fluorescence is maximum. Then, the position information acquisition unit 240 may calculate the distance from any one end of the detection device 100 in the three-dimensional space based on a camera parameter of the camera that serves as the sensor 400. The position information acquisition unit 240 may further use the shape of the detection device 100 measured in advance to calculate the distance from any one end of the detection device 100 in the three-dimensional space.
The detection unit 250 determines whether the portion of the image detection device 100 in the image where the intensity of fluorescence detected by the position information acquisition unit 240 is maximum is a portion where the intensity of fluorescence detected in the past is maximum in the history of measurement results. Specifically, the detection unit 250 detects, in the history of the measurement results, a portion from which a distance to the portion where the detected intensity of fluorescence is maximum (hereinafter, also referred to as maximum intensity portion) is equal to or less than a predetermined distance and in which the detected intensity of fluorescence is maximum (hereinafter, also referred to as past maximum intensity portion). The detection unit 250 stores a value representing the brightness of the maximum intensity portion in the measurement data storage unit 260 as the latest measurement result of the brightness of the detected past maximum intensity portion in the history of the measurement results. In a case where the past maximum intensity portion has not been detected, the detection unit 250 stores a value representing the brightness of the maximum intensity portion in the measurement data storage unit 260 as a new value of the brightness of the portion where the intensity of fluorescence is maximum in the history of the measurement results.
In a case where the detection device 100 is, for example, an optical fiber onto which the substance is applied and which leaks light, the light source 300 irradiates the applied substance with light via the detection device 100 that is the optical fiber.
In a case where the shape of the detection portion 110 represents the identification information of the detection device 102 including the detection portion 110, the detection portion 110 also operates as the identification information portion 140. In this case, the outline of the detection portion 110 may be drawn by using a substance that emits fluorescence when irradiated with light regardless of the presence or absence of the plant hormone. The shape of the detection portion 110 may be, for example, a predetermined figure, a character, or the like.
In this case, the position information acquisition unit 240 detects the region of the detection portion 110 from the image that is the measurement data. Note that, instead of the position information acquisition unit 240, the detection unit 250 may detect the region of the detection portion 110. The position information acquisition unit 240 recognizes the shape of the detection portion 110 detected. In this case, a shape that most matches the shape of the detection portion 110 among a plurality of predetermined shapes is estimated using any one of existing methods in which transformation such as affine transformation or perspective transformation and template matching are combined. The position information acquisition unit 240 may estimate the shape of the detection portion 110 in the three-dimensional space by using any one of existing methods using the camera parameter and the like. Then, the position information acquisition unit 240 may estimate a shape that most matches the estimated shape of the detection portion 110 among a plurality of predetermined shapes. The position information acquisition unit 240 holds in advance a combination of the shape of a figure and information indicating a place in which the detection device 102 including the detection portion 110 having the shape is installed, for each of a plurality of predetermined shapes. Then, the position information acquisition unit 240 determines that the place in which the detection device 102 is installed and which is combined with the shape that most matches the estimated shape of the detection portion 110 is the place in which the detection device 102 including the measured detection portion 110 is installed.
In a case where the shape of the detection portion 110 is a character, the position information acquisition unit 240 performs character recognition on a region including the region of the detection portion 110 by using any one of existing technologies for recognizing a character in a captured image. In this case, the position information acquisition unit 240 holds in advance a combination of a character and a place in which the detection device 102 including the detection portion 110 having the shape of the character is installed, for a plurality of predetermined characters. The position information acquisition unit 240 determines that the place in which the detection device 102 is installed and which is combined with the character recognized by the above-described character recognition is the place in which the detection device 102 including the detection portion 110 on which the above-described character recognition has been performed is installed.
Effects
The present example embodiment has the same effect as those of the fourth example embodiment. The reason is the same as the reason why the effects of the fourth example embodiment are exhibited.
Modified Example of Fifth Example Embodiment
The sensor 400 may be a photodetector having sensitivity to a wavelength of fluorescence emitted by the detection portion 110 irradiated with light. In this case, the measurement data is a value representing the brightness. Then, the detection unit 250 stores a value representing the brightness indicated by the measurement data acquired by the measurement data acquisition unit 230 in the measurement data storage unit 260.
Sixth Example Embodiment
FIG. 8 is a block diagram illustrating an example of a configuration of a terminal device 500 according to a sixth example embodiment of the present disclosure. The terminal device 500 of the present example embodiment is the same as the terminal device 500 according to the fifth example embodiment. The terminal device 500 of the present example embodiment is connected to the assistance device 201 according to the fifth example embodiment.
<Terminal Device 500>
The terminal device 500 includes an information receiving unit 510, a display control unit 520, and a display unit 530. The terminal device 500 can be implemented by, for example, a terminal device such as a smartphone, a tablet terminal, or a notebook personal computer.
The information receiving unit 510 receives the position information of the infected plant from the output unit 220 of the assistance device 201. The display control unit 520 controls the display unit 530 to display the received position information of the infected plant. The display unit 530 displays the position information of the infected plant. The display unit 530 is, for example, a display.
Modified Example of Sixth Example Embodiment
The terminal device 500 may be connected to the assistance device 200 according to the fourth example embodiment.
Other Example Embodiments
The assistance device 200 and the assistance device 201 described above can be implemented by a computer including a memory into which a program read from a storage medium is loaded and a processor that executes the program. The assistance device 200 and the assistance device 201 can also be implemented by dedicated hardware. The assistance device 200 and the assistance device 201 can also be implemented by a combination of the above-described computer and dedicated hardware.
FIG. 9 is a diagram illustrating an example of a hardware configuration of a computer 1000 that can implement the assistance devices 200 and 201 according to the above-described example embodiments. Referring to FIG. 9 , the computer 1000 includes a processor 1001, a memory 1002, a storage device 1003, and an input/output (I/O) interface 1004. In addition, the computer 1000 can access a storage medium 1005. The memory 1002 and the storage device 1003 are, for example, storage devices such as a random access memory (RAM) and a hard disk. The storage medium 1005 is, for example, a storage device such as a RAM or a hard disk, a read only memory (ROM), or a portable storage medium. The storage device 1003 may be the storage medium 1005. The processor 1001 can read and write data and programs from and in the memory 1002 and the storage device 1003. The processor 1001 can access, for example, the sensor 400 and the terminal device 500 via the I/O interface 1004. The processor 1001 can access the storage medium 1005. The storage medium 1005 stores a program for operating the computer 1000 as the assistance device according to the example embodiment of the present disclosure.
The processor 1001 loads, into the memory 1002, a program that is stored in the storage medium 1005 and causes the computer 1000 to operate as the assistance device according to the example embodiment of the present disclosure. Then, the processor 1001 executes the program loaded into the memory 1002, whereby the computer 1000 operates as the assistance device according to the example embodiment of the present disclosure.
The generation unit 210, the output unit 220, the measurement data acquisition unit 230, the position information acquisition unit 240, and the detection unit 250 can be implemented by, for example, the processor 1001 that executes the program loaded into the memory 1002. The measurement data storage unit 260 can be implemented by the memory 1002 included in the computer 1000 or the storage device 1003 such as a hard disk device. Some or all of the generation unit 210, the output unit 220, the measurement data acquisition unit 230, the position information acquisition unit 240, the detection unit 250, and the measurement data storage unit 260 can be implemented by a dedicated circuit.
Some or all of the above-described example embodiments can also be described as the following Supplementary Notes, but are not limited thereto.
[Supplementary Note 1]
A detection device including:
    • a detection portion that contains a substance that emits fluorescence in a case of being irradiated with light when reacting with a plant hormone.
      [Supplementary Note 2]
The detection device according to Supplementary Note 1, in which the substance emits the fluorescence with an intensity that is in proportion to an amount of the reacted plant hormone.
[Supplementary Note 3]
The detection device according to Supplementary Note 1 or 2, further including:
    • a reference portion that contains the substance; and
    • a shielding portion that is detachable and shields the reference portion from outside air.
      [Supplementary Note 4]
The detection device according to any one of Supplementary Notes 1 to 3, further including:
    • an identification information portion that holds identification information in a readable manner.
      [Supplementary Note 5]
The detection device according to any one of Supplementary Notes 1 to 4,
    • in which the substance includes at least one of a boron-oxygen compound having a structure represented by the following General Formula (5) as a receptor that selectively recognizes methyl salicylate, a rare earth compound as a receptor that selectively recognizes methyl salicylate, a zinc compound as a receptor that selectively recognizes methyl salicylate, or a hydrazine derivative as a receptor that selectively recognizes methyl jasmonate.
Figure US12298245-20250513-C00005

[Supplementary Note 6]
An assistance device including:
    • a generation unit that generates position information of a plant infected with a disease based on a captured image of a detection portion irradiated with light and a position of the detection portion, the detection portion containing a substance that emits fluorescence in a case of being irradiated with the light when reacting with a plant hormone; and an output unit that outputs the generated position information.
      [Supplementary Note 7]
The assistance device according to Supplementary Note 6,
    • in which the substance emits the fluorescence with an intensity that is in proportion to an amount of the reacted plant hormone, and
    • the generation unit generates the position information based on an amount of change in brightness of the detection portion in the captured image and the position of the detection portion.
      [Supplementary Note 8]
The assistance device according to Supplementary Note 6 or 7,
    • in which the captured image includes a region of the detection portion and a region of a reference portion that is imaged in a state where a shielding portion that is detachable and shields the reference portion containing the substance from outside air is removed from the reference portion, and
    • the generation unit generates the position information based on the amount of change in difference between the brightness of the detection portion and a brightness of the reference portion in the captured image.
      [Supplementary Note 9]
The assistance device according to any one of Supplementary Notes 6 to 8,
    • in which the generation unit generates, as the position information, a distribution of a degree of possibility that a plant infected with a disease exists in a place in which the detection portion is disposed.
      [Supplementary Note 10]
The assistance device according to any one of Supplementary Notes 6 to 8,
    • in which the substance includes at least one of a boron-oxygen compound having a structure represented by General Formula (1) as a receptor that selectively recognizes methyl salicylate, a rare earth compound as a receptor that selectively recognizes methyl salicylate, a zinc compound as a receptor that selectively recognizes methyl salicylate, or a hydrazine derivative as a receptor that selectively recognizes methyl jasmonate.
      [Supplementary Note 11]
An assistance system including:
    • the assistance device according to any one of Supplementary Notes 6 to 10;
    • a detection device including the detection portion;
    • an imaging device that captures the captured image; and
    • a terminal device that receives the position information and displays the position information.
      Supplementary Note 12
A terminal device including:
    • a receiving unit that receives position information of a plant infected with a disease, the position information being generated based on a captured image of a detection portion irradiated with light and a position of the detection portion, and the detection portion containing a substance that emits fluorescence in a case of being irradiated with the light when reacting with a plant hormone; and
    • a display unit that displays the received position information.
      [Supplementary Note 13]
The terminal device according to Supplementary Note 12, further including:
    • a receiving unit that receives the position information from an assistance device including a generation unit that generates the position information and an output unit that outputs the generated position information; and
    • a display control unit that controls the display unit to display the received position information.
      [Supplementary Note 14]
The terminal device according to Supplementary Note 13,
    • in which the substance emits the fluorescence with an intensity that is in proportion to an amount of the reacted plant hormone, and
    • the generation unit generates the position information based on the amount of change in brightness of the detection portion in the captured image and the position of the detection portion.
      [Supplementary Note 15]
The terminal device according to Supplementary Note 13 or 14,
    • in which the captured image includes a region of the detection portion and a region of a reference portion that is imaged in a state where a shielding portion that is detachable and shields the reference portion containing the substance from outside air is removed from the reference portion, and
    • the generation unit generates the position information based on the amount of change in difference between the brightness of the detection portion and a brightness of the reference portion in the captured image.
      [Supplementary Note 16]
The terminal device according to any one of Supplementary Notes 13 to 15,
    • in which the generation unit generates, as the position information, a distribution of a degree of possibility that a plant infected with a disease exists in a place in which the detection portion is disposed.
      [Supplementary Note 17]
The terminal device according to any one of Supplementary Notes 13 to 16,
    • in which the substance includes at least one of a boron-oxygen compound having a structure represented by General Formula (1) as a receptor that selectively recognizes methyl salicylate, a rare earth compound as a receptor that selectively recognizes methyl salicylate, a zinc compound as a receptor that selectively recognizes methyl salicylate, or a hydrazine derivative as a receptor that selectively recognizes methyl jasmonate.
      [Supplementary Note 18]
An assistance method including:
    • generating position information of a plant infected with a disease based on a captured image of a detection portion irradiated with light and a position of the detection portion, the detection portion containing a substance that emits fluorescence in a case of being irradiated with the light when reacting with a plant hormone; and
    • outputting the generated position information.
      [Supplementary Note 19]
The assistance method according to Supplementary Note 18,
    • in which the substance emits the fluorescence with an intensity that is in proportion to an amount of the reacted plant hormone, and
    • the assistance method further includes generating the position information based on the amount of change in brightness of the detection portion in the captured image and the position of the detection portion.
      [Supplementary Note 20]
The assistance method according to Supplementary Note 18 or 19,
    • in which the captured image includes a region of the detection portion and a region of a reference portion that is imaged in a state where a shielding portion that is detachable and shields the reference portion containing the substance from outside air is removed from the reference portion, and
    • the assistance method further includes generating the position information based on the amount of change in difference between the brightness of the detection portion and a brightness of the reference portion in the captured image.
      [Supplementary Note 21]
The assistance method according to any one of Supplementary Notes 18 to 20, further including:
    • generating, as the position information, a distribution of a degree of possibility that a plant infected with a disease exists in a place in which the detection portion is disposed.
      [Supplementary Note 22]
The assistance method according to any one of Supplementary Notes 18 to 21,
    • in which the substance includes at least one of a boron-oxygen compound having a structure represented by General Formula (1) as a receptor that selectively recognizes methyl salicylate, a rare earth compound as a receptor that selectively recognizes methyl salicylate, a zinc compound as a receptor that selectively recognizes methyl salicylate, or a hydrazine derivative as a receptor that selectively recognizes methyl jasmonate.
      [Supplementary Note 23]
A program for causing a computer to perform:
    • generation processing of generating position information of a plant infected with a disease based on a captured image of a detection portion irradiated with light and a position of the detection portion, the detection portion containing a substance that emits fluorescence in a case of being irradiated with the light when reacting with a plant hormone; and
    • output processing of outputting the generated position information.
      [Supplementary Note 24]
The program according to Supplementary Note 23,
    • in which the substance emits the fluorescence with an intensity that is in proportion to an amount of the reacted plant hormone, and
    • the generation processing includes generating the position information based on the amount of change in brightness of the detection portion in the captured image and the position of the detection portion.
      [Supplementary Note 25]
The program according to Supplementary Note 23 or 24,
    • in which the captured image includes a region of the detection portion and a region of a reference portion that is imaged in a state where a shielding portion that is detachable and shields the reference portion containing the substance from outside air is removed from the reference portion, and
    • the generation processing includes generating the position information based on the amount of change in difference between the brightness of the detection portion and a brightness of the reference portion in the captured image.
      [Supplementary Note 26]
The program according to any one of Supplementary Notes 23 to 25,
    • in which the generation processing includes generating, as the position information, a distribution of a degree of possibility that a plant infected with a disease exists in a place in which the detection portion is disposed.
      [Supplementary Note 27]
The program according to any one of Supplementary Notes 23 to 26,
    • in which the substance includes at least one of a boron-oxygen compound having a structure represented by General Formula (1) as a receptor that selectively recognizes methyl salicylate, a rare earth compound as a receptor that selectively recognizes methyl salicylate, a zinc compound as a receptor that selectively recognizes methyl salicylate, or a hydrazine derivative as a receptor that selectively recognizes methyl jasmonate.
In the technology described in International Publication No. WO 2021/111621, it is not possible to recognize the presence of a plant with an early-stage disease before the plant is spotted.
The present disclosure has an effect of enabling recognition of the presence of a plant with an early-stage disease.
While the present disclosure has been particularly shown and described with reference to exemplary embodiments thereof, the present disclosure is not limited to these embodiments. It will be understood by those of ordinary skill in the art that various changes in form and details may be made therein without departing from the spirit and scope of the present disclosure as defined by the claims.

Claims (15)

What is claimed is:
1. A detection device comprising:
a detection portion that contains a substance that emits fluorescence in a case of being irradiated with light when reacting with a plant hormone.
2. The detection device according to claim 1, wherein
the substance emits the fluorescence with an intensity that is in proportion to an amount of the reacted plant hormone.
3. The detection device according to claim 1, further comprising:
a reference portion that contains the substance; and
a shielding portion that is detachable and shields the reference portion from outside air.
4. The detection device according to claim 1, further comprising:
an identification information portion that holds identification information in a readable manner.
5. The detection device according to claim 1, wherein
the substance includes at least one of a boron-oxygen compound having a structure represented by General Formula (1) as a receptor that selectively recognizes methyl salicylate, a rare earth compound as a receptor that selectively recognizes methyl salicylate, a zinc compound as a receptor that selectively recognizes methyl salicylate, or a hydrazine derivative as a receptor that selectively recognizes methyl jasmonate
Figure US12298245-20250513-C00006
6. An assistance device comprising:
at least one memory configured to store instructions; and
at least one processor configured to execute the instructions to:
generate position information of a plant infected with a disease based on a captured image of a detection portion irradiated with light and a position of the detection portion, the detection portion containing a substance that emits fluorescence in a case of being irradiated with the light when reacting with a plant hormone; and
output the generated position information.
7. The assistance device according to claim 6, wherein
the captured image includes a region of the detection portion and a region of a reference portion that is imaged in a state where a shielding portion that is detachable and shields the reference portion containing the substance from outside air is removed from the reference portion, and
the at least one processor is further configured to generate the position information based on an amount of change in difference between a brightness of the detection portion and a brightness of the reference portion in the captured image.
8. The assistance device according to claim 6, wherein
the at least one processor is further configured to generate, as the position information, a distribution of a degree of possibility that a plant infected with a disease exists in a place in which the detection portion is disposed.
9. The assistance device according to claim 6, wherein
the substance includes at least one of a boron-oxygen compound having a structure represented by General Formula (1) as a receptor that selectively recognizes methyl salicylate, a rare earth compound as a receptor that selectively recognizes methyl salicylate, a zinc compound as a receptor that selectively recognizes methyl salicylate, or a hydrazine derivative as a receptor that selectively recognizes methyl jasmonate.
10. An assistance system including the assistance device according to claim 6, comprising:
a detection device including the detection portion;
an imaging device that captures the captured image; and
a terminal device that receives the position information and displays the position information.
11. An assistance method comprising:
generating position information of a plant infected with a disease based on a captured image of a detection portion irradiated with light and a position of the detection portion, the detection portion containing a substance that emits fluorescence in a case of being irradiated with the light when reacting with a plant hormone; and
outputting the generated position information.
12. The assistance method according to claim 11, wherein
the captured image includes a region of the detection portion and a region of a reference portion that is imaged in a state where a shielding portion that is detachable and shields the reference portion containing the substance from outside air is removed from the reference portion, and
the assistance method further comprises generating the position information based on an amount of change in difference between a brightness of the detection portion and a brightness of the reference portion in the captured image.
13. The assistance method according to claim 11, further comprising
generating, as the position information, a distribution of a degree of possibility that a plant infected with a disease exists in a place in which the detection portion is disposed.
14. The assistance method according to claim 11, wherein
the substance includes at least one of a boron-oxygen compound having a structure represented by General Formula (1) as a receptor that selectively recognizes methyl salicylate, a rare earth compound as a receptor that selectively recognizes methyl salicylate, a zinc compound as a receptor that selectively recognizes methyl salicylate, or a hydrazine derivative as a receptor that selectively recognizes methyl jasmonate.
15. The assistance method according to claim 11, further comprising:
capturing, using an imaging device, the captured image of the detection portion included in a detection device;
receiving, using a terminal device, the position information; and
displaying, using the terminal device, the position information.
US17/859,364 2021-07-27 2022-07-07 Detection device, assistance device and assistance method Active 2043-12-01 US12298245B2 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2021122182A JP2023018230A (en) 2021-07-27 2021-07-27 Device for detection, support device, support method and program
JP2021-122182 2021-07-27

Publications (2)

Publication Number Publication Date
US20230096759A1 US20230096759A1 (en) 2023-03-30
US12298245B2 true US12298245B2 (en) 2025-05-13

Family

ID=85158182

Family Applications (1)

Application Number Title Priority Date Filing Date
US17/859,364 Active 2043-12-01 US12298245B2 (en) 2021-07-27 2022-07-07 Detection device, assistance device and assistance method

Country Status (2)

Country Link
US (1) US12298245B2 (en)
JP (1) JP2023018230A (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021153513A1 (en) * 2020-01-27 2021-08-05 日本電気株式会社 Method for sensing plant hormone, and method for early detection of disease infection in plant using same
JP2023018230A (en) * 2021-07-27 2023-02-08 日本電気株式会社 Device for detection, support device, support method and program

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6711283B1 (en) * 2000-05-03 2004-03-23 Aperio Technologies, Inc. Fully automatic rapid microscope slide scanner
US20090286263A1 (en) * 2008-03-05 2009-11-19 United States Army As Represented By The Secretary Of The Army Skin tape stripping: a non-invasive diagnostic strategy for dermal exposure to cytotoxic compounds
US20210088444A1 (en) * 2018-03-07 2021-03-25 Konica Minolta, Inc. Image Processing Method, Image Processing Device, and Program
WO2021111621A1 (en) * 2019-12-06 2021-06-10 株式会社ナイルワークス Pathological diagnosis system for plants, pathological diagnosis method for plants, pathological diagnosis device for plants, and drone
US20230096759A1 (en) * 2021-07-27 2023-03-30 Nec Corporation Detection device, assistance device and assistance method
US20230292647A1 (en) * 2020-06-16 2023-09-21 Dark Horse Technologies Ltd System and Method for Crop Monitoring
US20240037749A1 (en) * 2020-12-08 2024-02-01 Terramera, Inc. Systems and methods for the improved detection of plants
US20240161928A1 (en) * 2022-11-15 2024-05-16 Quest Photonic Devices B.V. Computer based clinical decision support system and method for determining a classification of a lymphedema induced fluorescence pattern
US20240272416A1 (en) * 2021-07-28 2024-08-15 PAIGE.AI, Inc. Systems and methods for processing electronic images of pathology data and reviewing the pathology data
US12094199B2 (en) * 2021-03-31 2024-09-17 EarthSense, Inc. Agricultural analysis robotic systems and methods thereof

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP7084639B2 (en) * 2017-10-24 2022-06-15 国立大学法人 筑波大学 Plant stress detection method and photoprotein detection method in plants
US20190242861A1 (en) * 2018-02-08 2019-08-08 3M Innovative Properties Company Patterned chemical indicator
EP3918318A4 (en) * 2019-01-31 2023-03-01 C2Sense, Inc. Gas sensing identification
JP2021099259A (en) * 2019-12-23 2021-07-01 国立研究開発法人農業・食品産業技術総合研究機構 Method of quickly visualizing physiological effect of chemical substance applied to plant foliage
WO2021153513A1 (en) * 2020-01-27 2021-08-05 日本電気株式会社 Method for sensing plant hormone, and method for early detection of disease infection in plant using same
JP7400967B2 (en) * 2020-06-03 2023-12-19 日本電気株式会社 A method for sensing plant hormones using rare earth compounds, a sensor using the same, and a method for early detection of plant disease infection
US20240183779A1 (en) * 2021-03-20 2024-06-06 Nec Corporation Method for sensing methyl salicylate, methyl salicylate sensor, and method for detecting pathogen infection of plants

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6711283B1 (en) * 2000-05-03 2004-03-23 Aperio Technologies, Inc. Fully automatic rapid microscope slide scanner
US20090286263A1 (en) * 2008-03-05 2009-11-19 United States Army As Represented By The Secretary Of The Army Skin tape stripping: a non-invasive diagnostic strategy for dermal exposure to cytotoxic compounds
US20210088444A1 (en) * 2018-03-07 2021-03-25 Konica Minolta, Inc. Image Processing Method, Image Processing Device, and Program
WO2021111621A1 (en) * 2019-12-06 2021-06-10 株式会社ナイルワークス Pathological diagnosis system for plants, pathological diagnosis method for plants, pathological diagnosis device for plants, and drone
US20230292647A1 (en) * 2020-06-16 2023-09-21 Dark Horse Technologies Ltd System and Method for Crop Monitoring
US20240037749A1 (en) * 2020-12-08 2024-02-01 Terramera, Inc. Systems and methods for the improved detection of plants
US12094199B2 (en) * 2021-03-31 2024-09-17 EarthSense, Inc. Agricultural analysis robotic systems and methods thereof
US20230096759A1 (en) * 2021-07-27 2023-03-30 Nec Corporation Detection device, assistance device and assistance method
US20240272416A1 (en) * 2021-07-28 2024-08-15 PAIGE.AI, Inc. Systems and methods for processing electronic images of pathology data and reviewing the pathology data
US20240161928A1 (en) * 2022-11-15 2024-05-16 Quest Photonic Devices B.V. Computer based clinical decision support system and method for determining a classification of a lymphedema induced fluorescence pattern

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Real-time monitoring of plant stresses via chemiresistive profiling of leaf volatiles by a wearable sensor Li, Zheng et al. Matter, vol. 4, Issue 7, 2553-2570 (Year: 2021). *

Also Published As

Publication number Publication date
JP2023018230A (en) 2023-02-08
US20230096759A1 (en) 2023-03-30

Similar Documents

Publication Publication Date Title
US12298245B2 (en) Detection device, assistance device and assistance method
US12258242B2 (en) Systems and methods for verifying building material objects
EP3771198B1 (en) Target tracking method and device, movable platform and storage medium
US9470792B2 (en) Method and handheld distance measuring device for creating a spatial model
CN109212510A (en) Method and apparatus for measuring the angular resolution of multi-line laser radar
US10061994B2 (en) Driving-support-image generation device, driving-support-image display device, driving-support-image display system, and driving-support-image generation program
CN101151418A (en) Device for constructing weed map
WO2012037549A1 (en) Methods and apparatus for tracking motion and/or orientation of a marking device
CN118566995B (en) Mineral resource exploration methods, apparatus, equipment, and media based on subatomic fields
EP2752808A1 (en) Imaging device, work support program, information provision method, and information provision program
US20180068158A1 (en) Information processing device, information processing system, position reporting method, and program recording medium
US20200296684A1 (en) Terminal device for position measurement, computer program, and system
US9823354B2 (en) Illuminance measuring system
US20190133546A1 (en) System and method for locating a target subject
US20090084855A1 (en) Methods and Apparatus for Directing Bar Code Positioning for Imaging Scanning
JP2024023980A (en) Object management system, object management device, object management method and program
CN114503174A (en) Object recognition device, object recognition system, and object recognition method
CN111325045A (en) Bar code positioning system and bar code positioning method for scanner
WO2019001001A1 (en) Obstacle information acquisition apparatus and method
US11887250B2 (en) Guide display device and crane equipped with same
EP4129884A1 (en) Guide display device and crane provided with same
US20230097473A1 (en) Guide display device and crane equipped with same
US20230106308A1 (en) Guide display device and crane equipped with same
CN113407869A (en) Beacon labeling method, device, computer equipment and storage medium
CN219551512U (en) Soil and stone square quantity measuring device for airport construction

Legal Events

Date Code Title Description
AS Assignment

Owner name: NEC CORPORATION, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MAEDA, KATSUMI;REEL/FRAME:060430/0982

Effective date: 20220516

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STCF Information on status: patent grant

Free format text: PATENTED CASE