US20210066812A1 - System and method for radio frequency penetration imaging of an object - Google Patents

System and method for radio frequency penetration imaging of an object Download PDF

Info

Publication number
US20210066812A1
US20210066812A1 US17/096,119 US202017096119A US2021066812A1 US 20210066812 A1 US20210066812 A1 US 20210066812A1 US 202017096119 A US202017096119 A US 202017096119A US 2021066812 A1 US2021066812 A1 US 2021066812A1
Authority
US
United States
Prior art keywords
signals
target medium
medium
antenna array
reflected
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US17/096,119
Inventor
Damian HOFFMAN
Jonathan Rosenfeld
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Vayyar Imaging Ltd
Original Assignee
Vayyar Imaging Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Vayyar Imaging Ltd filed Critical Vayyar Imaging Ltd
Priority to US17/096,119 priority Critical patent/US20210066812A1/en
Assigned to VAYYAR IMAGING LTD reassignment VAYYAR IMAGING LTD ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ROSENFELD, JONATHAN
Assigned to VAYYAR IMAGING LTD reassignment VAYYAR IMAGING LTD ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: HOFFMAN, Damian
Publication of US20210066812A1 publication Critical patent/US20210066812A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01QANTENNAS, i.e. RADIO AERIALS
    • H01Q17/00Devices for absorbing waves radiated from an antenna; Combinations of such devices with active antenna elements or systems
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01QANTENNAS, i.e. RADIO AERIALS
    • H01Q1/00Details of, or arrangements associated with, antennas
    • H01Q1/12Supports; Mounting means
    • H01Q1/22Supports; Mounting means by structural association with other equipment or articles
    • H01Q1/24Supports; Mounting means by structural association with other equipment or articles with receiving set
    • H01Q1/241Supports; Mounting means by structural association with other equipment or articles with receiving set used in mobile communications, e.g. GSM
    • H01Q1/242Supports; Mounting means by structural association with other equipment or articles with receiving set used in mobile communications, e.g. GSM specially adapted for hand-held use
    • H01Q1/243Supports; Mounting means by structural association with other equipment or articles with receiving set used in mobile communications, e.g. GSM specially adapted for hand-held use with built-in antennas
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S13/00Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
    • G01S13/88Radar or analogous systems specially adapted for specific applications
    • G01S13/887Radar or analogous systems specially adapted for specific applications for detection of concealed objects, e.g. contraband or weapons
    • G01S13/888Radar or analogous systems specially adapted for specific applications for detection of concealed objects, e.g. contraband or weapons through wall detection
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S13/00Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
    • G01S13/88Radar or analogous systems specially adapted for specific applications
    • G01S13/89Radar or analogous systems specially adapted for specific applications for mapping or imaging
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01QANTENNAS, i.e. RADIO AERIALS
    • H01Q1/00Details of, or arrangements associated with, antennas
    • H01Q1/12Supports; Mounting means
    • H01Q1/22Supports; Mounting means by structural association with other equipment or articles
    • H01Q1/225Supports; Mounting means by structural association with other equipment or articles used in level-measurement devices, e.g. for level gauge measurement
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01QANTENNAS, i.e. RADIO AERIALS
    • H01Q21/00Antenna arrays or systems
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01QANTENNAS, i.e. RADIO AERIALS
    • H01Q21/00Antenna arrays or systems
    • H01Q21/06Arrays of individually energised antenna units similarly polarised and spaced apart
    • H01Q21/061Two dimensional planar arrays
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01QANTENNAS, i.e. RADIO AERIALS
    • H01Q5/00Arrangements for simultaneous operation of antennas on two or more different wavebands, e.g. dual-band or multi-band arrangements

Definitions

  • the present invention relates to a system and method for imaging an object or structure and more specifically, but not exclusively, to RF (Radio Frequency) penetration imaging of an object or structure within a medium.
  • RF Radio Frequency
  • Prior RF devices for imaging or scanning within a medium such as ground penetrating radars (GPR) typically assume a fixed, or minimal variation, distance between the imaging device antennas to the surface of the medium throughout the scanning or imaging process. Moreover, this distance is typically smaller than the transmission wavelength (i.e. sub-wavelength distance), i.e. the imaging device antennas are attached or in very close proximity to the surface of the medium.
  • GPR ground penetrating radars
  • Knowing where the signals were collected relative to the medium surface is critical for an aligned focusing of the signals as part of the imaging and/or medium scanning process.
  • prior devices can be less than ideal in at least some respects.
  • Prior devices having high imaging and accurate capabilities can be larger than ideal for use in many portable applications.
  • the cost of prior devices can be greater than would be ideal.
  • the prior devices can be somewhat bulky, difficult to activate requiring for example specific and constant distance from the medium, for example prior devices must be attached to the medium.
  • an improved imaging system that overcomes at least some of the above mentioned deficiencies of the prior devices would be beneficial.
  • such imaging device would be a compact, integrated with a consumer device such as a cellular telephone, sufficiently rugged and low in cost to be practical for end-user imaging of objects or structures within a medium and convenient to use.
  • Penetration RF imaging or ‘Standoff penetration RF imaging’ as used herein is defined as the process of imaging by an imaging system a medium internal structure or objects while not being in direct contact with its surface, e.g. at a distance larger than the wavelength used by the imaging device.
  • target medium as used herein is defined as an homogeneous or close to homogenous material in which scattering objects (i.e. objects or structures of different electromagnetic properties than the medium) may be present in and may be for example air, concrete, plaster, fluids, human tissues etc.
  • interim medium as used herein is defined as an homogeneous or close to homogeneous material medium located between the device and the target medium, of different electromagnetic properties that the target medium, such that the surface of the target medium is distinct from the interim medium, typically the interim medium being free of objects or structures, the device moving freely in this medium, the interim medium may be for example air, fluids, etc.
  • a RF (Radio Frequency) imaging system for imaging at least one object within a target medium
  • the system comprising a generation and reception unit configured to generate and receive a plurality of RF signals; an RF antenna array, the RF antenna array comprises a plurality of antennas, said plurality of antennas are configured to transmit the plurality of RF signals towards the target medium and receive a plurality of RF signals reflected from the target medium; a data acquisition unit configured to receive and store said plurality of reflected RF signals; and at least one processor unit, said at least one processor unit is configured to:estimate the distance between the surface of the target medium and the antenna array; calculate the delay between the plurality of transmitted signals and the plurality of signals reflected from the at least one object; estimate the location of the plurality of antennas at each transmitting time; and provide an image of the at least one object location and the intensity of the plurality of reflected RF signals.
  • a generation and reception unit configured to generate and receive a plurality of RF signals
  • an RF antenna array
  • the at least one processor is configured to analyze the angular orientation between the antenna array and the target medium surface.
  • the system comprises a display and wherein the image is displayed on said display.
  • an interim medium is located between the target medium and the antenna array of the imaging system and the antenna array is positioned or moved in a distance from the target medium.
  • the distance is constant or variable.
  • the distance is larger than a wavelength.
  • the interim medium or the target medium are selected from the group consisting of: air, concrete, plaster, fluids, human tissues.
  • the image is 2D (two dimension) or 3D (three dimension) image.
  • the RF signals are Ultra Wide Band signals.
  • the RF imaging system is portable.
  • the antenna array is configured to communicate with orientation or measurements devices, said devices are selected from the group consisting of: accelerometers, gyros, laser distance measurement devices.
  • a method for imaging at least one object within a target medium comprising: transmitting a plurality of RF (radio frequency) signals by an RF antenna array, said RF antenna array comprises a plurality of RF antennas; receiving at said RF antenna array a plurality of RF signals reflected from the at least one object; estimating the distance between the surface of the target medium and the antenna array; calculating the delay between the plurality of transmitted signals and the received signals reflected from the at least one object;
  • the method comprise estimating a location of the plurality of RF antennas at each transmitting time of said plurality of transmitted signals; and providing an image the at least one object location.
  • the method comprising providing the intensity of the reflected RF signals from the at least one object.
  • the meth comprising estimating the presence or lack of the at least one object.
  • an interim medium is located between the target medium and the antenna array of the imaging system and the antenna array is positioned or moved in a distance from the target medium.
  • the distance is constant or variable.
  • the angular orientation between the antenna array and the target medium surface is estimated from the received signals.
  • the distance between the surface of the target medium and the antenna array is estimated using the high frequency band of the plurality of received signals.
  • the delay is estimated from a subset of the plurality of received RF signals and/or using a dedicated frequency sub-band of the received signals.
  • the estimation of the location or orientation of the antenna array is improved using measurements from at least one or more accelerometers, gyros, cameras, laser devices.
  • Implementation of the method and/or system of embodiments of the invention can involve performing or completing selected tasks manually, automatically, or a combination thereof. Moreover, according to actual instrumentation and equipment of embodiments of the method and/or system of the invention, several selected tasks could be implemented by hardware, by software or by firmware or by a combination thereof using an operating system.
  • a data processor such as a computing platform for executing a plurality of instructions.
  • the data processor includes a volatile memory for storing instructions and/or data and/or a non-volatile storage, for example, a magnetic hard-disk and/or removable media, for storing instructions and/or data.
  • a network connection is provided as well.
  • a display and/or a user input device such as a keyboard or mouse are optionally provided as well.
  • FIG. 1 is a simplified schematic diagram illustrating a user scanning the surface of a target medium by a mobile device, in accordance with examples
  • FIG. 2 is a schematic view of an RF penetration imaging system, in accordance with examples
  • FIG. 3 is a simplified block diagram showing a system configuration for imaging a medium in accordance with examples
  • FIG. 4 is a flowchart illustrating a method for imaging a target medium and one or more objects located within a medium, in accordance with examples.
  • FIG. 5 is a simplified block diagram illustrating a rotation angle of the imaging in accordance with examples.
  • the present invention relates to a system, device and methods for RF imaging an object or structure within a medium and more specifically, but not exclusively, to an RF penetration imaging within a medium for example a standoff RF penetration imaging.
  • Embodiments of the disclosure provide a system and methods for imaging (e.g. 2D imaging or 3D imaging) of one or more objects (e.g. hidden objects to be imaged such as pipe or a miniaturized element) inside a target medium (e.g. solid medium such as a wall).
  • the system comprises an RF unit configured to transmit and receive RF signals which penetrate through the target medium (e.g. wall, fluids, etc.), one or more processing units configured to process the multiple RF signals reflected from the objects and/or the target medium surface and to provide an image of the hidden objects.
  • the visualization may be or may include a graphical visualization (e.g. rendered graphical visualization).
  • the graphical visualization may include an image such as a 2D or 3D image of the hidden objects inside the target medium comprising one or more of the object's parameters such as size, width volume etc., or indication of the presence or lack of objects within the medium.
  • an interim medium is located between the target medium and the imaging device and the antenna array may be positioned or moved at a distance from the medium.
  • a system and methods disclosed herein comprise scanning or imaging a medium regardless of the distance between the device (e.g. the device antennas) and the target medium surface, e.g. allowing for the distance from device antennas (e.g. the antenna array) to the target medium surface to change throughout the scanning process and not be limited to near field (e.g. sub-wavelength and/or comparable wavelength) distance, for example, attached or almost attached to the target medium or far-field (e.g. larger than wavelength) standoff distance.
  • near field e.g. sub-wavelength and/or comparable wavelength
  • received RF signals for example signals received by the system's receive antennas are combined in a coherent way by compensating for the occurred propagation delays within the medium. These delays can be estimated from a subset of received RF signals and/or for example using a dedicated frequency sub-band of the received signals.
  • system and method may be used in conjunction with additional orientation and/or distance measurement devices including for example accelerometers, gyros, laser distance measuring devices.
  • additional orientation and/or distance measurement devices including for example accelerometers, gyros, laser distance measuring devices.
  • FIG. 1 is a simplified schematic diagram showing a user 120 scanning the surface of a target medium such as a wall 140 by an imaging device 130 .
  • the imaging device 130 may be in communication or mounted possibly but not exclusively as an add-on, for example on a mobile device 100 , (e.g. mobile phone).
  • the target medium may be made of for example a solid element such as wood, stone, gypsum, concrete or non solid elements such as air, water, or a combination thereof.
  • the mobile device 100 may be for example a communication mobile device such as a mobile phone for example a smartphone or a tablet as will be illustrated specifically herein below in FIG. 2 .
  • the user 120 places the imaging device 130 in proximity to the wall for example in parallel to the wall (e.g. in front of the wall, few cm from the surface of the wall for example between 1-100 cm from the wall, or less). The distance may change during the scanning.
  • one or more hidden objects such as vertical rebars 122 and 123 embedded within the wall in front or in proximity to the device 130 may be imaged and displayed to the user in real time (or offline), for example on a display unit 110 of the device 100 .
  • the image may include a 3D image of the objects including specific information of the object's parameters such as size, orientation, radius, depth, and distance from the object to the imaging device.
  • a complete image of the wall 140 and the internal objects embedded in the wall 140 may be identified and imaged.
  • the image may be or may include a visualization such as a graphical visualization or a rendered graphical visualization of the object including the object's parameters such as size, orientation, distance from the medium etc.
  • a visualization such as a graphical visualization or a rendered graphical visualization of the object including the object's parameters such as size, orientation, distance from the medium etc.
  • the imaging device 130 is further configured to interface with the mobile device 100 computing platform.
  • the mobile device 100 may be for example a smartphone which may include a processor and software modules associated with the imaging device 130 which are executed by the mobile device processor.
  • the imaging device 130 may be attached to the mobile device 100 , and the imaging device 130 may be in communication to one or more processing units and/or a display unit of the mobile device 100 .
  • the user 120 may scan the surface of the wall by the imaging device 130 and the scanned data may be transmitted via wire or wireless connection to a database unit located for example in a cloud or in the mobile device 100 and the scanned data may be processed by one or more external processing units or internal processing unit of the imaging device 130 or the mobile device 100 .
  • the imaging device 130 may be a standalone device, for example a standalone portable device.
  • the scanned date (e.g. graphic visualization, parameters, or an image of the object) may be displayed to the user on headset (goggles) such as a VR (Virtual Reality) and/or AR (Augmented Reality) headset or on the mobile device such as smartphone.
  • headset goggles
  • VR Virtual Reality
  • AR Augmented Reality
  • the imaging device 130 and/or the mobile device 100 can have a size and weight such that the device can be held by a user with only one hand.
  • the imaging device 130 can have a size and weight such that the device can be portable.
  • the device can have a weight of about 100 gram (g), 200 g, 300 g, 400 g, 500 g, 600 g, 700 g, 800 g, 900 g, 1000 g.
  • the imaging device can have a weight that is between any of the two values given above.
  • the imaging device can have a weight greater than 1000 g.
  • the imaging device 130 shape can comprise a rectangular prism, cylinder, or other three-dimensional shape.
  • One or more of the components of the imaging device 130 can be powered by a battery.
  • FIG. 2 is block diagram illustrating the architecture of an RF penetration imaging system 200 .
  • the system comprises the principal functions of the above mentioned imaging device 130 and relevant functions of a mobile device 100 which may be attached to or in communication with it.
  • the imaging system 200 may include an RF signal generation and reception subsystem 204 , an antenna array subsystem 202 , a data acquisition subsystem 206 , and one or more data processors or processing units 208 , and additional sensors such as one or more accelerometers and/or gyroscopes 211 and/or one or more cameras 213 and/or a console 210 (for display storage and analysis).
  • additional sensors such as one or more accelerometers and/or gyroscopes 211 and/or one or more cameras 213 and/or a console 210 (for display storage and analysis).
  • the RF signal generation and reception (transmit-receive) subsystem 204 is responsible for generation of the microwave signals, coupling them to the antennas 202 a - 202 e , reception of the microwave signals from the antennas and converting them into a form suitable for acquisition.
  • the RF signals can be pulse signals, stepped-frequency signals, chirp signals and the like.
  • the generation circuitry can involve oscillators, synthesizers, mixers, or it can be based on pulse oriented circuits such as logic gates or step-recovery diodes. For example these signals may be microwave signals in the UWB band 3, 4, 5, 6, 7, 8, 9 or 10 GHz (having a wavelength of 10-3 cm correspondingly in air).
  • the conversion process can include down conversion, sampling, and the like.
  • the conversion process typically includes averaging in the form of low-pass filtering, to improve the signal-to-noise ratios and to allow for lower sampling rates.
  • the RF signal generation and reception subsystem 204 can perform transmission and reception with multiple antennas at a time or select one transmit and one receive antenna at a time, according to a tradeoff between complexity and acquisition time.
  • the RF signal generation and reception subsystem 204 may comprise a single RF Integrated Circuit (RFIC) or several RFICs.
  • RFIC RF Integrated Circuit
  • the imaging system 200 may include one or more antennas such as antenna array 202 .
  • the antenna array 202 may include multiple antennas 202 a - 202 e typically between a few and several dozen (for example 30) antennas.
  • the antennas can be of many types known in the art, such as printed antennas, waveguide antennas, dipole antennas or “Vivaldi” broadband antennas.
  • the antenna array can be linear or two-dimensional, flat or conformal to the region of interest.
  • the antenna array 202 comprises a number of antennas for each frequency range, or a single antenna for each frequency range, or a single antenna supporting the entire frequency range.
  • the array may include 20 antennas where 5 antennas are dedicated for each out of the 4 frequency bands.
  • the antenna array 202 may be an array of flat broadband antennas, for example spiral shaped antennas.
  • the antenna array 202 may include a layer of matching material for improved coupling of the antenna radiation to the materials or objects under test.
  • the unique and optimized shape of the antenna array enables their use in limited sized mobile devices, such as a thin, small-sized smart phone or tablet.
  • the data acquisition subsystem 206 collects and digitizes the signals from the RF signal generation and reception subsystem 204 while tagging the signals according to the antenna combination used and the time at which the signals were collected.
  • the data acquisition subsystem will typically include analog-to-digital (A/D) converters and data buffers, but it may include additional functions such as signal averaging, correlation of waveforms with templates or converting signals between frequency and time domain.
  • A/D analog-to-digital
  • the data acquisition subsystem 206 may include a Radio Frequency Signals Imaging device (RFSMU) such as a Vector Network Analyzer (VNA) for measuring the received/reflected signals.
  • RFSMU Radio Frequency Signals Imaging device
  • VNA Vector Network Analyzer
  • the data processing unit 208 is responsible for converting the collected signals into a set of responses characterizing the target objects, and performing the algorithms for converting the sets of responses, for example providing the location of the target objects, the intensity of the signal reflection from the objects.
  • Example of algorithms for converting the sets of responses may be for example Delay and Sum (DAS) algorithm, PFA (polar format SAR algorithm), RMA (Range Migration Algorithm).
  • DAS Delay and Sum
  • PFA polar format SAR algorithm
  • RMA Range Migration Algorithm
  • the imaging system 200 may include accelerometers and/or gyros and/or laser distance measuring devices 211 to fine tune and give additional data in respect to the movement or the distance of the imaging system 200 .
  • the imaging system 200 may include a camera 213 to obtain the system relative location or movement in respect to a reference location, as will be illustrated in details hereinabove.
  • Such additional measurements can be used as prior information to improve the location and orientation calculation using the signals received and transmitted by the antennas, or can be used to improve the calculation of location and orientation after it has been calculated from the signals received and transmitted by the antennas.
  • the improvement can be achieved using a variety of methods, such as Kalman Filtering, Particle Filtering, Bayesian Filtering or other methods.
  • a final step in the process includes making use of the resulting parameters or image, either in the form of visualization, display, storage, archiving, or input to feature detection algorithms.
  • This step is exemplified in FIG. 2 as console 210 .
  • the console for example may be part of a mobile device for example implemented as a handheld computer such as a mobile phone, laptop computer or a desktop computer with appropriate application software.
  • an imaging device 230 which is part of the RF penetration imaging system 200 , may comprise the RF transmit-receive subsystem 204 and the antenna array subsystem 202 as shown in FIG. 2 , while a mobile device 220 (e.g. mobile phone) may comprise some or all the remaining functional blocks of 200 .
  • the imaging device may be connected to the mobile device 220 , for example through a dedicated connection such a USB connection, wireless connection or any connection known in the art. Following the connection of the imaging device 230 to the mobile device 220 , the imaging device 230 may utilize the mobile device's own data acquisition, data processing display, and storage and analysis subsystems.
  • data acquisition subsystem 206 , data processing unit 208 as well as measuring devices 211 and 213 may also be part of the imaging device 230 (in addition to RF transmit-receive subsystem 204 and the antenna array subsystem 202 ).
  • the console 210 may also be integrated into the imaging device 230 , thus getting a complete stand alone system.
  • the console 210 is typically implemented as a handheld computer such as a mobile phone or a desktop computer with appropriate application software. It is stressed that other embodiments comprising alternative arrangements of the subsystems between the imaging device 230 the RF penetration imaging system 200 and the mobile device are applicable.
  • FIG. 3 is a simplified block diagram side view description showing a system configuration 300 for imaging a target medium 340 and one or more objects (e.g. one or more scattering objects 350 ) within the target medium in accordance with embodiments.
  • objects e.g. one or more scattering objects 350
  • the system 300 comprises an RF penetration imaging system 320 which may comprise one of the units and subsystems illustrated in system 200 of FIG. 2 . Only the antenna array subsystem 302 , the RF generation and reception subsystem 304 and the data acquisition subsystem 306 and processing unit 308 are shown in FIG. 3 .
  • the imaging device is positioned and/or moved at a distance 370 from the target medium surface 333 .
  • the distance 370 may the distance between the antenna array 302 and the target medium outer surface 333 . In some cases the distance 370 may be constant or variable as the user scans and image the target medium.
  • the antenna array 302 transmits one or more RF signals generated by the RF signal generation and reception subsystem 304 towards the target medium 340 .
  • the antenna array 302 When the antenna array 302 is not attached to the target medium 340 (or in very close proximity with it), e.g. at a standoff distance, the RF signals propagate through an interim medium 330 .
  • the interim medium 330 may be or may include air or other intermediate matched medium located between the antenna array 302 and a target medium 340 .
  • FIG. 4 is a flowchart 400 illustrating a method for imaging one or more objects located within the target medium, in accordance with embodiments.
  • Step 410 comprises transmitting one or more RF signals by one or more antennas of the antenna array 302 and step 415 comprises receiving at the antenna array one or more signals reflected by the outer surface 333 of the target medium and/or by the objects 350 within the target medium.
  • Step 420 comprises estimating the distance 370 between the antenna array and the outer surface 333 of the target medium, for example estimating the distance between the surface of the target medium 340 and the antenna array 302 .
  • This distance estimation may be provided, for example, using the high frequency band of the RF signals. For example (for a case where the entire band is 3-10 GHz) the higher frequency band of the signals used for this estimation may be 6-10 GHz.
  • step 420 includes evaluation of the antenna array angular orientation in respect to the target medium surface.
  • the angular orientation may include the rotation angle ⁇ of the antenna array 502 in respect to the surface of the target medium 540 .
  • Step 430 comprises measuring the propagation delay between the transmitted signals and received signals reflected by the scattering objects 350 inside the target medium 340 .
  • the different electromagnetic properties of the interim and target medium materials and the distance measured in accordance with step 420 are taken into account to separate the delay incurred within the target medium (i.e. the propagation delay from the outer surface of the target medium to the object) out of the overall propagation delay.
  • Step 440 comprises estimating the location of each antenna in space (e.g. antenna 302 a - d ) relative to the target medium (e.g. the surface 333 of the target medium 340 ) during the scanning process (e.g. while transmitting and receiving the signals affected by the medium). This estimation can be performed using, but not limited to, inertial measurements device on the antennas or the signals themselves.
  • Step 450 comprises providing an image of the scattered objects 350 location and the intensity of their reflected RF signal, using one or more methods, such as, but not limited to, DAS (delay and sum), PFA (polar format SAR algorithm), RMA (Range Migration Algorithm).
  • DAS delay and sum
  • PFA polar format SAR algorithm
  • RMA Range Migration Algorithm
  • Step 460 comprises estimating the presence or lack of scattered objects inside the target medium according to the image provided in step 450 .
  • This embodiment can use RF signals such as UWB (Ultra Wide Band) signals, covering a large frequency range (e.g. 3-10 GHz).
  • the higher frequency band of the signals e.g. 6-10 GHz
  • the whole frequency band of the RF signals and the lower frequency band in particular are used to estimate the overall delay (Step 430 ), with a larger (but sufficient) accuracy. It is noted that at lower frequencies the signals incur a smaller propagation path loss in the target medium.
  • the processing unit may be a digital processing device including one or more hardware central processing units (CPU) that carry out the device's functions.
  • the digital processing device further comprises an operating system configured to perform executable instructions.
  • the digital processing device is optionally connected a computer network.
  • the digital processing device is optionally connected to the Internet such that it accesses the World Wide Web.
  • the digital processing device is optionally connected to a cloud computing infrastructure.
  • the digital processing device is optionally connected to an intranet.
  • the digital processing device is optionally connected to a data storage device.
  • suitable digital processing devices include, by way of non-limiting examples, server computers, desktop computers, laptop computers, notebook computers, sub-notebook computers, netbook computers, netpad computers, set-top computers, handheld computers, Internet appliances, mobile smartphones, tablet computers, personal digital assistants, video game consoles, and vehicles.
  • server computers desktop computers, laptop computers, notebook computers, sub-notebook computers, netbook computers, netpad computers, set-top computers, handheld computers, Internet appliances, mobile smartphones, tablet computers, personal digital assistants, video game consoles, and vehicles.
  • smartphones are suitable for use in the system described herein.
  • Suitable tablet computers include those with booklet, slate, and convertible configurations, known to those of skill in the art.
  • the digital processing device includes an operating system configured to perform executable instructions.
  • the operating system is, for example, software, including programs and data, which manages the device's hardware and provides services for execution of applications.
  • suitable server operating systems include, by way of non-limiting examples, FreeBSD, OpenBSD, NetBSD®, Linux, Apple® Mac OS X Server®, Oracle® Solaris®, Windows Server®, and Novell® NetWare®.
  • suitable personal computer operating systems include, by way of non-limiting examples, Microsoft® Windows®, Apple® Mac OS X®, UNIX®, and UNIX-like operating systems such as GNU/Linux®.
  • the operating system is provided by cloud computing.
  • suitable mobile smart phone operating systems include, by way of non-limiting examples, Nokia® Symbian® OS, Apple® iOS®, Research In Motion® BlackBerry OS®, Google® Android®, Microsoft® Windows Phone® OS, Microsoft® Windows Mobile® OS, Linux®, and Palm® WebOS®.
  • the device includes a storage and/or memory device.
  • the storage and/or memory device is one or more physical apparatuses used to store data or programs on a temporary or permanent basis.
  • the device is volatile memory and requires power to maintain stored information.
  • the device is non-volatile memory and retains stored information when the digital processing device is not powered.
  • the non-volatile memory comprises flash memory.
  • the non-volatile memory comprises dynamic random-access memory (DRAM).
  • the non-volatile memory comprises ferroelectric random access memory (FRAM).
  • the non-volatile memory comprises phase-change random access memory (PRAM).
  • the device is a storage device including, by way of non-limiting examples, CD-ROMs, DVDs, flash memory devices, magnetic disk drives, magnetic tapes drives, optical disk drives, and cloud computing based storage.
  • the storage and/or memory device is a combination of devices such as those disclosed herein.
  • the digital processing device includes a display to send visual information to a user.
  • the display is a cathode ray tube (CRT).
  • the display is a liquid crystal display (LCD).
  • the display is a thin film transistor liquid crystal display (TFT-LCD).
  • the display is an organic light emitting diode (OLED) display.
  • OLED organic light emitting diode
  • on OLED display is a passive-matrix OLED (PMOLED) or active-matrix OLED (AMOLED) display.
  • the display is a plasma display.
  • the display is a video projector.
  • the display is a combination of devices such as those disclosed herein.
  • the digital processing device includes an input device to receive information from a user.
  • the input device is a keyboard.
  • the input device is a pointing device including, by way of non-limiting examples, a mouse, trackball, track pad, joystick, game controller, or stylus.
  • the input device is a touch screen or a multi-touch screen.
  • the input device is a microphone to capture voice or other sound input.
  • the input device is a video camera to capture motion or visual input.
  • the input device is a combination of devices such as those disclosed herein.
  • the system disclosed herein includes one or more non-transitory computer readable storage media encoded with a program including instructions executable by the operating system of an optionally networked digital processing device.
  • a computer readable storage medium is a tangible component of a digital processing device.
  • a computer readable storage medium is optionally removable from a digital processing device.
  • a computer readable storage medium includes, by way of non-limiting examples, CD-ROMs, DVDs, flash memory devices, solid state memory, magnetic disk drives, magnetic tape drives, optical disk drives, cloud computing systems and services, and the like.
  • the program and instructions are permanently, substantially permanently, semi-permanently, or non-transitorily encoded on the media.
  • the system disclosed herein includes at least one computer program, or use of the same.
  • a computer program includes a sequence of instructions, executable in the digital processing device's CPU, written to perform a specified task.
  • Computer readable instructions may be implemented as program modules, such as functions, objects, Application Programming Interfaces (APIs), data structures, and the like, that perform particular tasks or implement particular abstract data types.
  • APIs Application Programming Interfaces
  • a computer program comprises one sequence of instructions. In some embodiments, a computer program comprises a plurality of sequences of instructions. In some embodiments, a computer program is provided from one location. In other embodiments, a computer program is provided from a plurality of locations. In various embodiments, a computer program includes one or more software modules. In various embodiments, a computer program includes, in part or in whole, one or more web applications, one or more mobile applications, one or more standalone applications, one or more web browser plug-ins, extensions, add-ins, or add-ons, or combinations thereof.
  • a computer program includes a mobile application provided to a mobile digital processing device.
  • the mobile application is provided to a mobile digital processing device at the time it is manufactured.
  • the mobile application is provided to a mobile digital processing device via the computer network described herein.
  • a mobile application is created by techniques known to those of skill in the art using hardware, languages, and development environments known to the art. Those of skill in the art will recognize that mobile applications are written in several languages. Suitable programming languages include, by way of non-limiting examples, C, C++, C#, Objective-C, JavaTM, Javascript, Pascal, Object Pascal, PythonTM, Ruby, VB.NET, WML, and XHTML/HTML with or without CSS, or combinations thereof.
  • Suitable mobile application development environments are available from several sources.
  • Commercially available development environments include, by way of non-limiting examples, AirplaySDK, alcheMo, Appcelerator®, Celsius, Bedrock, Flash Lite, .NET Compact Framework, Rhomobile, and WorkLight Mobile Platform.
  • Other development environments are available without cost including, by way of non-limiting examples, Lazarus, MobiFlex, MoSync, and Phonegap.
  • mobile device manufacturers distribute software developer kits including, by way of non-limiting examples, iPhone and iPad (iOS) SDK, AndroidTM SDK, BlackBerry® SDK, BREW SDK, Palm® OS SDK, Symbian SDK, webOS SDK, and Windows® Mobile SDK.
  • the system disclosed herein includes software, server, and/or database modules, or use of the same.
  • software modules are created by techniques known to those of skill in the art using machines, software, and languages known to the art.
  • the software modules disclosed herein are implemented in a multitude of ways.
  • a software module comprises a file, a section of code, a programming object, a programming structure, or combinations thereof.
  • a software module comprises a plurality of files, a plurality of sections of code, a plurality of programming objects, a plurality of programming structures, or combinations thereof.
  • the one or more software modules comprise, by way of non-limiting examples, a web application, a mobile application, and a standalone application.
  • software modules are in one computer program or application. In other embodiments, software modules are in more than one computer program or application. In some embodiments, software modules are hosted on one machine. In other embodiments, software modules are hosted on more than one machine. In further embodiments, software modules are hosted on cloud computing platforms. In some embodiments, software modules are hosted on one or more machines in one location. In other embodiments, software modules are hosted on one or more machines in more than one location.
  • the system disclosed herein includes one or more databases, or use of the same.
  • suitable databases include, by way of non-limiting examples, relational databases, non-relational databases, object oriented databases, object databases, entity-relationship model databases, associative databases, and XML, databases.
  • a database is internet-based.
  • a database is web-based.
  • a database is cloud computing-based.
  • a database is based on one or more local computer storage devices.
  • Methods of the present invention may be implemented by performing or completing manually, automatically, or a combination thereof, selected steps or tasks.
  • the present invention may be implemented in the testing or practice with methods and materials equivalent or similar to those described herein.

Landscapes

  • Engineering & Computer Science (AREA)
  • Remote Sensing (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Physics & Mathematics (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Electromagnetism (AREA)
  • General Physics & Mathematics (AREA)
  • Radar Systems Or Details Thereof (AREA)

Abstract

A system and methods for RF (Radio Frequency) penetration imaging of one or more objects in a medium, the system including a generation and reception subsystem configured to generate and receive a plurality of RF signals, an RF antenna array including a plurality of antennas, the antennas being configured to transmit the RF signals towards the medium and receive a plurality of RF signals reflected from the medium, a data acquisition subsystem configured to receive and store the reflected RF signals, and a processor configured to estimate the distance between the surface of the target medium and the antenna array, the delay between the transmitted signals and the plurality of signals reflected from the object using a dedicated frequency sub-band of the received signals, the location of the antennas at each transmitting time, and determine whether there is an object within the medium.

Description

    FIELD OF THE INVENTION
  • The present invention relates to a system and method for imaging an object or structure and more specifically, but not exclusively, to RF (Radio Frequency) penetration imaging of an object or structure within a medium.
  • BACKGROUND OF THE INVENTION
  • Prior RF devices for imaging or scanning within a medium, such as ground penetrating radars (GPR), typically assume a fixed, or minimal variation, distance between the imaging device antennas to the surface of the medium throughout the scanning or imaging process. Moreover, this distance is typically smaller than the transmission wavelength (i.e. sub-wavelength distance), i.e. the imaging device antennas are attached or in very close proximity to the surface of the medium.
  • These two assumptions, separately, and together, play an important role in the ability to form a coherent and accurate image for the analysis of the underlying structures and objects within the medium. Amongst such image techniques one finds, radiograms, migration techniques.
  • Knowing where the signals were collected relative to the medium surface is critical for an aligned focusing of the signals as part of the imaging and/or medium scanning process.
  • Additionally, the prior devices can be less than ideal in at least some respects. Prior devices having high imaging and accurate capabilities can be larger than ideal for use in many portable applications. Also, the cost of prior devices can be greater than would be ideal. The prior devices can be somewhat bulky, difficult to activate requiring for example specific and constant distance from the medium, for example prior devices must be attached to the medium.
  • In light of the above, an improved imaging system that overcomes at least some of the above mentioned deficiencies of the prior devices would be beneficial. Ideally such imaging device would be a compact, integrated with a consumer device such as a cellular telephone, sufficiently rugged and low in cost to be practical for end-user imaging of objects or structures within a medium and convenient to use.
  • SUMMARY OF INVENTION
  • Prior to the Summary of the invention being set forth, it may be helpful to set forth definitions of certain terms that will be used hereinafter.
  • The term ‘Penetration RF imaging’ or ‘Standoff penetration RF imaging’ as used herein is defined as the process of imaging by an imaging system a medium internal structure or objects while not being in direct contact with its surface, e.g. at a distance larger than the wavelength used by the imaging device.
  • The term ‘target medium’ as used herein is defined as an homogeneous or close to homogenous material in which scattering objects (i.e. objects or structures of different electromagnetic properties than the medium) may be present in and may be for example air, concrete, plaster, fluids, human tissues etc.
  • The term ‘interim medium’ as used herein is defined as an homogeneous or close to homogeneous material medium located between the device and the target medium, of different electromagnetic properties that the target medium, such that the surface of the target medium is distinct from the interim medium, typically the interim medium being free of objects or structures, the device moving freely in this medium, the interim medium may be for example air, fluids, etc.
  • According to a first aspect of the invention there is provided a RF (Radio Frequency) imaging system, for imaging at least one object within a target medium, the system comprising a generation and reception unit configured to generate and receive a plurality of RF signals; an RF antenna array, the RF antenna array comprises a plurality of antennas, said plurality of antennas are configured to transmit the plurality of RF signals towards the target medium and receive a plurality of RF signals reflected from the target medium; a data acquisition unit configured to receive and store said plurality of reflected RF signals; and at least one processor unit, said at least one processor unit is configured to:estimate the distance between the surface of the target medium and the antenna array; calculate the delay between the plurality of transmitted signals and the plurality of signals reflected from the at least one object; estimate the location of the plurality of antennas at each transmitting time; and provide an image of the at least one object location and the intensity of the plurality of reflected RF signals.
  • In an embodiment, the at least one processor is configured to analyze the angular orientation between the antenna array and the target medium surface.
  • In an embodiment, the system comprises a display and wherein the image is displayed on said display.
  • In an embodiment, an interim medium is located between the target medium and the antenna array of the imaging system and the antenna array is positioned or moved in a distance from the target medium.
  • In an embodiment, the distance is constant or variable.
  • In an embodiment, the distance is larger than a wavelength.
  • In an embodiment, the interim medium or the target medium are selected from the group consisting of: air, concrete, plaster, fluids, human tissues.
  • In an embodiment, the image is 2D (two dimension) or 3D (three dimension) image.
  • In an embodiment, the RF signals are Ultra Wide Band signals.
  • In an embodiment, the RF imaging system is portable.
  • In an embodiment, the antenna array is configured to communicate with orientation or measurements devices, said devices are selected from the group consisting of: accelerometers, gyros, laser distance measurement devices.
  • According to a second aspect there is provided a method for imaging at least one object within a target medium, the method comprising: transmitting a plurality of RF (radio frequency) signals by an RF antenna array, said RF antenna array comprises a plurality of RF antennas; receiving at said RF antenna array a plurality of RF signals reflected from the at least one object; estimating the distance between the surface of the target medium and the antenna array; calculating the delay between the plurality of transmitted signals and the received signals reflected from the at least one object;
  • In an embodiment, the method comprise estimating a location of the plurality of RF antennas at each transmitting time of said plurality of transmitted signals; and providing an image the at least one object location.
  • In an embodiment, the method comprising providing the intensity of the reflected RF signals from the at least one object.
  • In an embodiment, the meth comprising estimating the presence or lack of the at least one object.
  • In an embodiment, an interim medium is located between the target medium and the antenna array of the imaging system and the antenna array is positioned or moved in a distance from the target medium.
  • In an embodiment, the distance is constant or variable.
  • In an embodiment, the angular orientation between the antenna array and the target medium surface is estimated from the received signals.
  • In an embodiment, the distance between the surface of the target medium and the antenna array is estimated using the high frequency band of the plurality of received signals.
  • In an embodiment, the delay is estimated from a subset of the plurality of received RF signals and/or using a dedicated frequency sub-band of the received signals.
  • In an embodiment, the estimation of the location or orientation of the antenna array is improved using measurements from at least one or more accelerometers, gyros, cameras, laser devices.
  • Unless otherwise defined, all technical and/or scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which the invention pertains. Although methods and materials similar or equivalent to those described herein can be used in the practice or testing of embodiments of the invention, exemplary methods and/or materials are described below. In case of conflict, the patent specification, including definitions, will control. In addition, the materials, methods, and examples are illustrative only and are not intended to be necessarily limiting.
  • Implementation of the method and/or system of embodiments of the invention can involve performing or completing selected tasks manually, automatically, or a combination thereof. Moreover, according to actual instrumentation and equipment of embodiments of the method and/or system of the invention, several selected tasks could be implemented by hardware, by software or by firmware or by a combination thereof using an operating system.
  • For example, hardware for performing selected tasks, according to embodiments of the invention, could be implemented as a chip or a circuit. As software, selected tasks according to embodiments of the invention could be implemented as a plurality of software instructions being executed by a computer using any suitable operating system. In an exemplary embodiment of the invention, one or more tasks according to exemplary embodiments of method and/or system as described herein, are performed by a data processor, such as a computing platform for executing a plurality of instructions. Optionally, the data processor includes a volatile memory for storing instructions and/or data and/or a non-volatile storage, for example, a magnetic hard-disk and/or removable media, for storing instructions and/or data. Optionally, a network connection is provided as well. A display and/or a user input device such as a keyboard or mouse are optionally provided as well.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The subject matter disclosed may best be understood by reference to the following detailed description when read with the accompanying drawings in which:
  • FIG. 1 is a simplified schematic diagram illustrating a user scanning the surface of a target medium by a mobile device, in accordance with examples;
  • FIG. 2 is a schematic view of an RF penetration imaging system, in accordance with examples;
  • FIG. 3 is a simplified block diagram showing a system configuration for imaging a medium in accordance with examples;
  • FIG. 4 is a flowchart illustrating a method for imaging a target medium and one or more objects located within a medium, in accordance with examples; and
  • FIG. 5 is a simplified block diagram illustrating a rotation angle of the imaging in accordance with examples.
  • DETAILED DESCRIPTION OF THE INVENTION
  • The present invention relates to a system, device and methods for RF imaging an object or structure within a medium and more specifically, but not exclusively, to an RF penetration imaging within a medium for example a standoff RF penetration imaging.
  • Embodiments of the disclosure provide a system and methods for imaging (e.g. 2D imaging or 3D imaging) of one or more objects (e.g. hidden objects to be imaged such as pipe or a miniaturized element) inside a target medium (e.g. solid medium such as a wall). The system comprises an RF unit configured to transmit and receive RF signals which penetrate through the target medium (e.g. wall, fluids, etc.), one or more processing units configured to process the multiple RF signals reflected from the objects and/or the target medium surface and to provide an image of the hidden objects. The visualization may be or may include a graphical visualization (e.g. rendered graphical visualization). For example, the graphical visualization may include an image such as a 2D or 3D image of the hidden objects inside the target medium comprising one or more of the object's parameters such as size, width volume etc., or indication of the presence or lack of objects within the medium. In some cases, an interim medium is located between the target medium and the imaging device and the antenna array may be positioned or moved at a distance from the medium.
  • Specifically, a system and methods disclosed herein comprise scanning or imaging a medium regardless of the distance between the device (e.g. the device antennas) and the target medium surface, e.g. allowing for the distance from device antennas (e.g. the antenna array) to the target medium surface to change throughout the scanning process and not be limited to near field (e.g. sub-wavelength and/or comparable wavelength) distance, for example, attached or almost attached to the target medium or far-field (e.g. larger than wavelength) standoff distance.
  • In other words, the two prior art assumptions mentioned herein above, i.e. requiring constant distance between the imaging device antennas to the surface of the target medium throughout the scanning or imaging process and/or a distance smaller than the antenna wavelength, are eliminated according to the present invention embodiments.
  • In some cases, received RF signals, for example signals received by the system's receive antennas are combined in a coherent way by compensating for the occurred propagation delays within the medium. These delays can be estimated from a subset of received RF signals and/or for example using a dedicated frequency sub-band of the received signals.
  • According to some embodiments the system and method may be used in conjunction with additional orientation and/or distance measurement devices including for example accelerometers, gyros, laser distance measuring devices.
  • Before explaining at least one embodiment of the invention in detail, it is to be understood that the invention is not necessarily limited in its application to the details of construction and the arrangement of the components and/or methods set forth in the following description and/or illustrated in the drawings and/or the examples. The invention is capable of other embodiments or of being practiced or carried out in various ways.
  • Referring now to the drawings, FIG. 1 is a simplified schematic diagram showing a user 120 scanning the surface of a target medium such as a wall 140 by an imaging device 130. In some embodiments the imaging device 130 may be in communication or mounted possibly but not exclusively as an add-on, for example on a mobile device 100, (e.g. mobile phone). The target medium may be made of for example a solid element such as wood, stone, gypsum, concrete or non solid elements such as air, water, or a combination thereof. The mobile device 100 may be for example a communication mobile device such as a mobile phone for example a smartphone or a tablet as will be illustrated specifically herein below in FIG. 2.
  • In operation, the user 120 places the imaging device 130 in proximity to the wall for example in parallel to the wall (e.g. in front of the wall, few cm from the surface of the wall for example between 1-100 cm from the wall, or less). The distance may change during the scanning. Once the user 120 scans the wall, one or more hidden objects such as vertical rebars 122 and 123 embedded within the wall in front or in proximity to the device 130 may be imaged and displayed to the user in real time (or offline), for example on a display unit 110 of the device 100. In some cases the image may include a 3D image of the objects including specific information of the object's parameters such as size, orientation, radius, depth, and distance from the object to the imaging device. In some cases a complete image of the wall 140 and the internal objects embedded in the wall 140 may be identified and imaged.
  • In some cases, the image may be or may include a visualization such as a graphical visualization or a rendered graphical visualization of the object including the object's parameters such as size, orientation, distance from the medium etc.
  • According to some embodiments the imaging device 130 is further configured to interface with the mobile device 100 computing platform. The mobile device 100 may be for example a smartphone which may include a processor and software modules associated with the imaging device 130 which are executed by the mobile device processor.
  • For example, the imaging device 130 may be attached to the mobile device 100, and the imaging device 130 may be in communication to one or more processing units and/or a display unit of the mobile device 100.
  • In some case, the user 120 may scan the surface of the wall by the imaging device 130 and the scanned data may be transmitted via wire or wireless connection to a database unit located for example in a cloud or in the mobile device 100 and the scanned data may be processed by one or more external processing units or internal processing unit of the imaging device 130 or the mobile device 100.
  • According to some embodiments, the imaging device 130 may be a standalone device, for example a standalone portable device.
  • In some cases the scanned date (e.g. graphic visualization, parameters, or an image of the object) may be displayed to the user on headset (goggles) such as a VR (Virtual Reality) and/or AR (Augmented Reality) headset or on the mobile device such as smartphone.
  • The imaging device 130 and/or the mobile device 100 can have a size and weight such that the device can be held by a user with only one hand. The imaging device 130 can have a size and weight such that the device can be portable. The device can have a weight of about 100 gram (g), 200 g, 300 g, 400 g, 500 g, 600 g, 700 g, 800 g, 900 g, 1000 g. The imaging device can have a weight that is between any of the two values given above. The imaging device can have a weight greater than 1000 g.
  • The imaging device 130 shape can comprise a rectangular prism, cylinder, or other three-dimensional shape.
  • One or more of the components of the imaging device 130 can be powered by a battery.
  • FIG. 2 is block diagram illustrating the architecture of an RF penetration imaging system 200. The system comprises the principal functions of the above mentioned imaging device 130 and relevant functions of a mobile device 100 which may be attached to or in communication with it.
  • The imaging system 200 may include an RF signal generation and reception subsystem 204, an antenna array subsystem 202, a data acquisition subsystem 206, and one or more data processors or processing units 208, and additional sensors such as one or more accelerometers and/or gyroscopes 211 and/or one or more cameras 213 and/or a console 210 (for display storage and analysis).
  • The RF signal generation and reception (transmit-receive) subsystem 204 is responsible for generation of the microwave signals, coupling them to the antennas 202 a-202 e, reception of the microwave signals from the antennas and converting them into a form suitable for acquisition. The RF signals can be pulse signals, stepped-frequency signals, chirp signals and the like. The generation circuitry can involve oscillators, synthesizers, mixers, or it can be based on pulse oriented circuits such as logic gates or step-recovery diodes. For example these signals may be microwave signals in the UWB band 3, 4, 5, 6, 7, 8, 9 or 10 GHz (having a wavelength of 10-3 cm correspondingly in air).
  • The conversion process can include down conversion, sampling, and the like. The conversion process typically includes averaging in the form of low-pass filtering, to improve the signal-to-noise ratios and to allow for lower sampling rates. The RF signal generation and reception subsystem 204 can perform transmission and reception with multiple antennas at a time or select one transmit and one receive antenna at a time, according to a tradeoff between complexity and acquisition time. In an embodiment, the RF signal generation and reception subsystem 204 may comprise a single RF Integrated Circuit (RFIC) or several RFICs.
  • The imaging system 200 may include one or more antennas such as antenna array 202. For example the antenna array 202 may include multiple antennas 202 a-202 e typically between a few and several dozen (for example 30) antennas. The antennas can be of many types known in the art, such as printed antennas, waveguide antennas, dipole antennas or “Vivaldi” broadband antennas. The antenna array can be linear or two-dimensional, flat or conformal to the region of interest.
  • According to one embodiment, the antenna array 202 comprises a number of antennas for each frequency range, or a single antenna for each frequency range, or a single antenna supporting the entire frequency range. According to another example, the array may include 20 antennas where 5 antennas are dedicated for each out of the 4 frequency bands.
  • According to some embodiment, the antenna array 202 may be an array of flat broadband antennas, for example spiral shaped antennas. The antenna array 202 may include a layer of matching material for improved coupling of the antenna radiation to the materials or objects under test. The unique and optimized shape of the antenna array enables their use in limited sized mobile devices, such as a thin, small-sized smart phone or tablet.
  • The data acquisition subsystem 206 collects and digitizes the signals from the RF signal generation and reception subsystem 204 while tagging the signals according to the antenna combination used and the time at which the signals were collected. The data acquisition subsystem will typically include analog-to-digital (A/D) converters and data buffers, but it may include additional functions such as signal averaging, correlation of waveforms with templates or converting signals between frequency and time domain.
  • The data acquisition subsystem 206 may include a Radio Frequency Signals Imaging device (RFSMU) such as a Vector Network Analyzer (VNA) for measuring the received/reflected signals.
  • The data processing unit 208 is responsible for converting the collected signals into a set of responses characterizing the target objects, and performing the algorithms for converting the sets of responses, for example providing the location of the target objects, the intensity of the signal reflection from the objects.
  • Example of algorithms for converting the sets of responses may be for example Delay and Sum (DAS) algorithm, PFA (polar format SAR algorithm), RMA (Range Migration Algorithm).
  • According to some embodiments, the imaging system 200 may include accelerometers and/or gyros and/or laser distance measuring devices 211 to fine tune and give additional data in respect to the movement or the distance of the imaging system 200.
  • Additionally, the imaging system 200 may include a camera 213 to obtain the system relative location or movement in respect to a reference location, as will be illustrated in details hereinabove.
  • Such additional measurements can be used as prior information to improve the location and orientation calculation using the signals received and transmitted by the antennas, or can be used to improve the calculation of location and orientation after it has been calculated from the signals received and transmitted by the antennas. The improvement can be achieved using a variety of methods, such as Kalman Filtering, Particle Filtering, Bayesian Filtering or other methods. Once the location and orientation information of the antennas array were calculated, it can be used in the imaging process as displacement information between measurements, for example in migration algorithms.
  • A final step in the process includes making use of the resulting parameters or image, either in the form of visualization, display, storage, archiving, or input to feature detection algorithms. This step is exemplified in FIG. 2 as console 210. The console for example may be part of a mobile device for example implemented as a handheld computer such as a mobile phone, laptop computer or a desktop computer with appropriate application software.
  • According to one embodiment of this invention, an imaging device 230, which is part of the RF penetration imaging system 200, may comprise the RF transmit-receive subsystem 204 and the antenna array subsystem 202 as shown in FIG. 2, while a mobile device 220 (e.g. mobile phone) may comprise some or all the remaining functional blocks of 200. The imaging device may be connected to the mobile device 220, for example through a dedicated connection such a USB connection, wireless connection or any connection known in the art. Following the connection of the imaging device 230 to the mobile device 220, the imaging device 230 may utilize the mobile device's own data acquisition, data processing display, and storage and analysis subsystems.
  • According to another embodiment, data acquisition subsystem 206, data processing unit 208 as well as measuring devices 211 and 213 may also be part of the imaging device 230 (in addition to RF transmit-receive subsystem 204 and the antenna array subsystem 202). The console 210 may also be integrated into the imaging device 230, thus getting a complete stand alone system. Alternatively, the console 210 is typically implemented as a handheld computer such as a mobile phone or a desktop computer with appropriate application software. It is stressed that other embodiments comprising alternative arrangements of the subsystems between the imaging device 230 the RF penetration imaging system 200 and the mobile device are applicable.
  • FIG. 3 is a simplified block diagram side view description showing a system configuration 300 for imaging a target medium 340 and one or more objects (e.g. one or more scattering objects 350) within the target medium in accordance with embodiments.
  • The system 300 comprises an RF penetration imaging system 320 which may comprise one of the units and subsystems illustrated in system 200 of FIG. 2. Only the antenna array subsystem 302, the RF generation and reception subsystem 304 and the data acquisition subsystem 306 and processing unit 308 are shown in FIG. 3.
  • In operation, the imaging device is positioned and/or moved at a distance 370 from the target medium surface 333. The distance 370 may the distance between the antenna array 302 and the target medium outer surface 333. In some cases the distance 370 may be constant or variable as the user scans and image the target medium. During the scanning process the antenna array 302 transmits one or more RF signals generated by the RF signal generation and reception subsystem 304 towards the target medium 340. When the antenna array 302 is not attached to the target medium 340 (or in very close proximity with it), e.g. at a standoff distance, the RF signals propagate through an interim medium 330. In some instances, the interim medium 330 may be or may include air or other intermediate matched medium located between the antenna array 302 and a target medium 340.
  • FIG. 4 is a flowchart 400 illustrating a method for imaging one or more objects located within the target medium, in accordance with embodiments.
  • Step 410 comprises transmitting one or more RF signals by one or more antennas of the antenna array 302 and step 415 comprises receiving at the antenna array one or more signals reflected by the outer surface 333 of the target medium and/or by the objects 350 within the target medium.
  • Step 420 comprises estimating the distance 370 between the antenna array and the outer surface 333 of the target medium, for example estimating the distance between the surface of the target medium 340 and the antenna array 302. This distance estimation may be provided, for example, using the high frequency band of the RF signals. For example (for a case where the entire band is 3-10 GHz) the higher frequency band of the signals used for this estimation may be 6-10 GHz. In addition, step 420 includes evaluation of the antenna array angular orientation in respect to the target medium surface. For example as illustrated in FIG. 5, the angular orientation may include the rotation angle α of the antenna array 502 in respect to the surface of the target medium 540.
  • Step 430 comprises measuring the propagation delay between the transmitted signals and received signals reflected by the scattering objects 350 inside the target medium 340. The different electromagnetic properties of the interim and target medium materials and the distance measured in accordance with step 420 are taken into account to separate the delay incurred within the target medium (i.e. the propagation delay from the outer surface of the target medium to the object) out of the overall propagation delay.
  • Step 440 comprises estimating the location of each antenna in space (e.g. antenna 302 a-d) relative to the target medium (e.g. the surface 333 of the target medium 340) during the scanning process (e.g. while transmitting and receiving the signals affected by the medium). This estimation can be performed using, but not limited to, inertial measurements device on the antennas or the signals themselves.
  • Step 450 comprises providing an image of the scattered objects 350 location and the intensity of their reflected RF signal, using one or more methods, such as, but not limited to, DAS (delay and sum), PFA (polar format SAR algorithm), RMA (Range Migration Algorithm).
  • Step 460 comprises estimating the presence or lack of scattered objects inside the target medium according to the image provided in step 450.
  • This embodiment can use RF signals such as UWB (Ultra Wide Band) signals, covering a large frequency range (e.g. 3-10 GHz). The higher frequency band of the signals (e.g. 6-10 GHz) enable increased accuracy in estimating the distance between the antenna array and the surface 333 of the target medium (Step 420). It is noted that typically the propagation path loss in the interim medium is small in spite of the high frequencies used. The whole frequency band of the RF signals and the lower frequency band in particular, are used to estimate the overall delay (Step 430), with a larger (but sufficient) accuracy. It is noted that at lower frequencies the signals incur a smaller propagation path loss in the target medium.
  • In further embodiments, the processing unit may be a digital processing device including one or more hardware central processing units (CPU) that carry out the device's functions. In still further embodiments, the digital processing device further comprises an operating system configured to perform executable instructions. In some embodiments, the digital processing device is optionally connected a computer network. In further embodiments, the digital processing device is optionally connected to the Internet such that it accesses the World Wide Web. In still further embodiments, the digital processing device is optionally connected to a cloud computing infrastructure. In other embodiments, the digital processing device is optionally connected to an intranet. In other embodiments, the digital processing device is optionally connected to a data storage device.
  • In accordance with the description herein, suitable digital processing devices include, by way of non-limiting examples, server computers, desktop computers, laptop computers, notebook computers, sub-notebook computers, netbook computers, netpad computers, set-top computers, handheld computers, Internet appliances, mobile smartphones, tablet computers, personal digital assistants, video game consoles, and vehicles. Those of skill in the art will recognize that many smartphones are suitable for use in the system described herein. Those of skill in the art will also recognize that select televisions with optional computer network connectivity are suitable for use in the system described herein. Suitable tablet computers include those with booklet, slate, and convertible configurations, known to those of skill in the art.
  • In some embodiments, the digital processing device includes an operating system configured to perform executable instructions. The operating system is, for example, software, including programs and data, which manages the device's hardware and provides services for execution of applications. Those of skill in the art will recognize that suitable server operating systems include, by way of non-limiting examples, FreeBSD, OpenBSD, NetBSD®, Linux, Apple® Mac OS X Server®, Oracle® Solaris®, Windows Server®, and Novell® NetWare®. Those of skill in the art will recognize that suitable personal computer operating systems include, by way of non-limiting examples, Microsoft® Windows®, Apple® Mac OS X®, UNIX®, and UNIX-like operating systems such as GNU/Linux®. In some embodiments, the operating system is provided by cloud computing. Those of skill in the art will also recognize that suitable mobile smart phone operating systems include, by way of non-limiting examples, Nokia® Symbian® OS, Apple® iOS®, Research In Motion® BlackBerry OS®, Google® Android®, Microsoft® Windows Phone® OS, Microsoft® Windows Mobile® OS, Linux®, and Palm® WebOS®.
  • In some embodiments, the device includes a storage and/or memory device. The storage and/or memory device is one or more physical apparatuses used to store data or programs on a temporary or permanent basis. In some embodiments, the device is volatile memory and requires power to maintain stored information. In some embodiments, the device is non-volatile memory and retains stored information when the digital processing device is not powered. In further embodiments, the non-volatile memory comprises flash memory. In some embodiments, the non-volatile memory comprises dynamic random-access memory (DRAM). In some embodiments, the non-volatile memory comprises ferroelectric random access memory (FRAM). In some embodiments, the non-volatile memory comprises phase-change random access memory (PRAM). In other embodiments, the device is a storage device including, by way of non-limiting examples, CD-ROMs, DVDs, flash memory devices, magnetic disk drives, magnetic tapes drives, optical disk drives, and cloud computing based storage. In further embodiments, the storage and/or memory device is a combination of devices such as those disclosed herein.
  • In some embodiments, the digital processing device includes a display to send visual information to a user. In some embodiments, the display is a cathode ray tube (CRT). In some embodiments, the display is a liquid crystal display (LCD). In further embodiments, the display is a thin film transistor liquid crystal display (TFT-LCD). In some embodiments, the display is an organic light emitting diode (OLED) display. In various further embodiments, on OLED display is a passive-matrix OLED (PMOLED) or active-matrix OLED (AMOLED) display. In some embodiments, the display is a plasma display. In other embodiments, the display is a video projector. In still further embodiments, the display is a combination of devices such as those disclosed herein.
  • In some embodiments, the digital processing device includes an input device to receive information from a user. In some embodiments, the input device is a keyboard. In some embodiments, the input device is a pointing device including, by way of non-limiting examples, a mouse, trackball, track pad, joystick, game controller, or stylus. In some embodiments, the input device is a touch screen or a multi-touch screen. In other embodiments, the input device is a microphone to capture voice or other sound input. In other embodiments, the input device is a video camera to capture motion or visual input. In still further embodiments, the input device is a combination of devices such as those disclosed herein.
  • In some embodiments, the system disclosed herein includes one or more non-transitory computer readable storage media encoded with a program including instructions executable by the operating system of an optionally networked digital processing device. In further embodiments, a computer readable storage medium is a tangible component of a digital processing device. In still further embodiments, a computer readable storage medium is optionally removable from a digital processing device.
  • In some embodiments, a computer readable storage medium includes, by way of non-limiting examples, CD-ROMs, DVDs, flash memory devices, solid state memory, magnetic disk drives, magnetic tape drives, optical disk drives, cloud computing systems and services, and the like. In some cases, the program and instructions are permanently, substantially permanently, semi-permanently, or non-transitorily encoded on the media. In some embodiments, the system disclosed herein includes at least one computer program, or use of the same. A computer program includes a sequence of instructions, executable in the digital processing device's CPU, written to perform a specified task. Computer readable instructions may be implemented as program modules, such as functions, objects, Application Programming Interfaces (APIs), data structures, and the like, that perform particular tasks or implement particular abstract data types. In light of the disclosure provided herein, those of skill in the art will recognize that a computer program may be written in various versions of various languages.
  • The functionality of the computer readable instructions may be combined or distributed as desired in various environments. In some embodiments, a computer program comprises one sequence of instructions. In some embodiments, a computer program comprises a plurality of sequences of instructions. In some embodiments, a computer program is provided from one location. In other embodiments, a computer program is provided from a plurality of locations. In various embodiments, a computer program includes one or more software modules. In various embodiments, a computer program includes, in part or in whole, one or more web applications, one or more mobile applications, one or more standalone applications, one or more web browser plug-ins, extensions, add-ins, or add-ons, or combinations thereof.
  • In some embodiments, a computer program includes a mobile application provided to a mobile digital processing device. In some embodiments, the mobile application is provided to a mobile digital processing device at the time it is manufactured. In other embodiments, the mobile application is provided to a mobile digital processing device via the computer network described herein.
  • In view of the disclosure provided herein, a mobile application is created by techniques known to those of skill in the art using hardware, languages, and development environments known to the art. Those of skill in the art will recognize that mobile applications are written in several languages. Suitable programming languages include, by way of non-limiting examples, C, C++, C#, Objective-C, Java™, Javascript, Pascal, Object Pascal, Python™, Ruby, VB.NET, WML, and XHTML/HTML with or without CSS, or combinations thereof.
  • Suitable mobile application development environments are available from several sources. Commercially available development environments include, by way of non-limiting examples, AirplaySDK, alcheMo, Appcelerator®, Celsius, Bedrock, Flash Lite, .NET Compact Framework, Rhomobile, and WorkLight Mobile Platform. Other development environments are available without cost including, by way of non-limiting examples, Lazarus, MobiFlex, MoSync, and Phonegap. Also, mobile device manufacturers distribute software developer kits including, by way of non-limiting examples, iPhone and iPad (iOS) SDK, Android™ SDK, BlackBerry® SDK, BREW SDK, Palm® OS SDK, Symbian SDK, webOS SDK, and Windows® Mobile SDK.
  • Those of skill in the art will recognize that several commercial forums are available for distribution of mobile applications including, by way of non-limiting examples, Apple® App Store, Android™ Market, BlackBerry® App World, App Store for Palm devices, App Catalog for webOS, Windows® Marketplace for Mobile, Ovi Store for Nokia® devices, Samsung® Apps, and Nintendo® DSi Shop.
  • In some embodiments, the system disclosed herein includes software, server, and/or database modules, or use of the same. In view of the disclosure provided herein, software modules are created by techniques known to those of skill in the art using machines, software, and languages known to the art. The software modules disclosed herein are implemented in a multitude of ways. In various embodiments, a software module comprises a file, a section of code, a programming object, a programming structure, or combinations thereof. In further various embodiments, a software module comprises a plurality of files, a plurality of sections of code, a plurality of programming objects, a plurality of programming structures, or combinations thereof. In various embodiments, the one or more software modules comprise, by way of non-limiting examples, a web application, a mobile application, and a standalone application. In some embodiments, software modules are in one computer program or application. In other embodiments, software modules are in more than one computer program or application. In some embodiments, software modules are hosted on one machine. In other embodiments, software modules are hosted on more than one machine. In further embodiments, software modules are hosted on cloud computing platforms. In some embodiments, software modules are hosted on one or more machines in one location. In other embodiments, software modules are hosted on one or more machines in more than one location.
  • In some embodiments, the system disclosed herein includes one or more databases, or use of the same. In view of the disclosure provided herein, those of skill in the art will recognize that many databases are suitable for storage and retrieval of information as described herein. In various embodiments, suitable databases include, by way of non-limiting examples, relational databases, non-relational databases, object oriented databases, object databases, entity-relationship model databases, associative databases, and XML, databases. In some embodiments, a database is internet-based. In further embodiments, a database is web-based. In still further embodiments, a database is cloud computing-based. In other embodiments, a database is based on one or more local computer storage devices.
  • In the above description, an embodiment is an example or implementation of the inventions. The various appearances of “one embodiment,” “an embodiment” or “some embodiments” do not necessarily all refer to the same embodiments.
  • Although various features of the invention may be described in the context of a single embodiment, the features may also be provided separately or in any suitable combination. Conversely, although the invention may be described herein in the context of separate embodiments for clarity, the invention may also be implemented in a single embodiment.
  • Reference in the specification to “some embodiments”, “an embodiment”, “one embodiment” or “other embodiments” means that a particular feature, structure, or characteristic described in connection with the embodiments is included in at least some embodiments, but not necessarily all embodiments, of the inventions.
  • It is to be understood that the phraseology and terminology employed herein is not to be construed as limiting and are for descriptive purpose only.
  • The principles and uses of the teachings of the present invention may be better understood with reference to the accompanying description, figures and examples.
  • It is to be understood that the details set forth herein do not construe a limitation to an application of the invention.
  • Furthermore, it is to be understood that the invention can be carried out or practiced in various ways and that the invention can be implemented in embodiments other than the ones outlined in the description above.
  • It is to be understood that the terms “including”, “comprising”, “consisting” and grammatical variants thereof do not preclude the addition of one or more components, features, steps, or integers or groups thereof and that the terms are to be construed as specifying components, features, steps or integers.
  • If the specification or claims refer to “an additional” element, that does not preclude there being more than one of the additional element.
  • It is to be understood that where the claims or specification refer to “a” or “an” element, such reference is not be construed that there is only one of that element.
  • It is to be understood that where the specification states that a component, feature, structure, or characteristic “may”, “might”, “can” or “could” be included, that particular component, feature, structure, or characteristic is not required to be included.
  • Where applicable, although state diagrams, flow diagrams or both may be used to describe embodiments, the invention is not limited to those diagrams or to the corresponding descriptions. For example, flow need not move through each illustrated box or state, or in exactly the same order as illustrated and described.
  • Methods of the present invention may be implemented by performing or completing manually, automatically, or a combination thereof, selected steps or tasks.
  • The descriptions, examples, methods and materials presented in the claims and the specification are not to be construed as limiting but rather as illustrative only.
  • Meanings of technical and scientific terms used herein are to be commonly understood as by one of ordinary skill in the art to which the invention belongs, unless otherwise defined.
  • The present invention may be implemented in the testing or practice with methods and materials equivalent or similar to those described herein.
  • While the invention has been described with respect to a limited number of embodiments, these should not be construed as limitations on the scope of the invention, but rather as exemplifications of some of the preferred embodiments. Other possible variations, modifications, and applications are also within the scope of the invention. Accordingly, the scope of the invention should not be limited by what has thus far been described, but by the appended claims and their legal equivalents.
  • All publications, patents and patent applications mentioned in this specification are herein incorporated in their entirety by reference into the specification, to the same extent as if each individual publication, patent or patent application was specifically and individually indicated to be incorporated herein by reference. In addition, citation or identification of any reference in this application shall not be construed as an admission that such reference is available as prior art to the present invention. To the extent that section headings are used, they should not be construed as necessarily limiting.

Claims (20)

What is claimed is:
1. A method for determining whether there is at least one object within a target medium, the method comprising:
transmitting a plurality of RF signals by a radio frequency (RF) antenna array to the target medium, wherein said RF antenna array comprises a plurality of RF antennas;
receiving a plurality of reflected RF signals from the target medium;
combining coherently the reflected RF signals by compensating for the occurred propagation delays within the target medium;
estimating the propagation delays, wherein the delays are estimated from a subset of received RF signals; and
determining based on said estimation whether there is at least one object within the target medium.
2. The method of claim 1, wherein the delay is estimated using a dedicated frequency sub-band of the reflected RF signals.
3. The method of claim 1, comprising providing an image of the at least one object location.
4. The method of claim 3, wherein the image is a two dimensional (2D) or a three dimensional (3D) image.
5. The method of claim 1 comprising providing the intensity of the reflected RF signal of the at least one object.
6. The method of claim 1 wherein said RF antenna array is positioned away at a distance from the target medium surface.
7. The method of claim 6, wherein the distance is constant or variable.
8. The method of claim 1, comprising estimating the angular orientation between the RF array and the target medium surface.
9. The method of claim 6, wherein the distance between the surface of the target medium and the RF array is estimated using the high frequency band of the RF reflected signals.
10. The method of claim 1 comprising scanning the target medium using a mobile device, wherein said mobile device comprising said RF antenna array.
11. A radio frequency (RF) system for determining whether there is at least one object within a target medium, the system comprising
a generation and reception subsystem configured to generate and receive a plurality of RF signals;
an RF antenna array, the RF antenna array comprises a plurality of RF antennas configured to transmit the plurality of RF signals towards the target medium and receive a plurality of RF signals reflected from the target medium;
a data acquisition subsystem configured to receive and store said plurality of reflected RF signals; and
at least one processor, said at least one processor is configured to:
combine coherently the reflected RF signals by compensating for the occurred propagation delays within the target medium;
estimate the propagation delays, wherein the delays are estimated from a subset of received RF signals; and
determine based on said estimation whether there is at least one object within the target medium.
12. The system of claim 11, wherein the at least one processor is further configured to provide an image of the at least one object location.
13. The system of claim 12, wherein the system comprises a display and wherein the image is displayed on said display.
14. The system of claim 11, wherein the at least one processor is further configured to provide the intensity of the plurality of reflected RF signals.
15. The system of claim 11, wherein said at least one processor is configured to analyze the angular orientation between the antenna array and the target medium surface.
16. The system of claim 11, wherein an interim medium is located between the target medium and the antenna array of the imaging system and the antenna array is positioned or moved in a distance from the target medium.
17. The system of claim 11, wherein the distance is constant or variable.
18. The system of claim 16, wherein an interim medium is located between the target medium and the antenna array of the imaging system and the RF array is positioned or moved in the distance from the target medium.
19. The system of claim 16, wherein the interim medium or the target medium is selected from the group consisting of: air, concrete, plaster, fluids, human tissues.
20. The system of claim 10, wherein the RF signals are Ultra Wide Band signals.
US17/096,119 2015-08-02 2020-11-12 System and method for radio frequency penetration imaging of an object Abandoned US20210066812A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US17/096,119 US20210066812A1 (en) 2015-08-02 2020-11-12 System and method for radio frequency penetration imaging of an object

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
US201562200075P 2015-08-02 2015-08-02
US15/226,048 US10290948B2 (en) 2015-08-02 2016-08-02 System and method for radio frequency penetration imaging of an object
US16/402,319 US10566699B2 (en) 2015-08-02 2019-05-03 System and method for radio frequency penetration imaging of an object
US16/749,393 US10854990B2 (en) 2015-08-02 2020-01-22 System and method for radio frequency penetration imaging of an object
US17/096,119 US20210066812A1 (en) 2015-08-02 2020-11-12 System and method for radio frequency penetration imaging of an object

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US16/749,393 Continuation US10854990B2 (en) 2015-08-02 2020-01-22 System and method for radio frequency penetration imaging of an object

Publications (1)

Publication Number Publication Date
US20210066812A1 true US20210066812A1 (en) 2021-03-04

Family

ID=57883750

Family Applications (4)

Application Number Title Priority Date Filing Date
US15/226,048 Active 2036-12-06 US10290948B2 (en) 2015-08-02 2016-08-02 System and method for radio frequency penetration imaging of an object
US16/402,319 Active US10566699B2 (en) 2015-08-02 2019-05-03 System and method for radio frequency penetration imaging of an object
US16/749,393 Active US10854990B2 (en) 2015-08-02 2020-01-22 System and method for radio frequency penetration imaging of an object
US17/096,119 Abandoned US20210066812A1 (en) 2015-08-02 2020-11-12 System and method for radio frequency penetration imaging of an object

Family Applications Before (3)

Application Number Title Priority Date Filing Date
US15/226,048 Active 2036-12-06 US10290948B2 (en) 2015-08-02 2016-08-02 System and method for radio frequency penetration imaging of an object
US16/402,319 Active US10566699B2 (en) 2015-08-02 2019-05-03 System and method for radio frequency penetration imaging of an object
US16/749,393 Active US10854990B2 (en) 2015-08-02 2020-01-22 System and method for radio frequency penetration imaging of an object

Country Status (1)

Country Link
US (4) US10290948B2 (en)

Families Citing this family (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11918330B2 (en) 2017-03-08 2024-03-05 Praesidium, Inc. Home occupant detection and monitoring system
US10989806B2 (en) 2017-03-08 2021-04-27 Praesidium, Inc. Home occupant detection and monitoring system
US11239570B2 (en) 2018-10-16 2022-02-01 Hook'd WiFi Inc. Wireless communications system with scalable architecture
US11183773B2 (en) 2018-10-16 2021-11-23 Hook'd WiFi Inc. Configurable communication system using stacked antennas
WO2020219967A1 (en) * 2019-04-25 2020-10-29 Hook'd WiFi Inc. Configurable communication system using stacked antennas
KR20210155946A (en) * 2020-06-17 2021-12-24 삼성전자주식회사 Electronic device for performing positioning and method thereof
US20220027628A1 (en) * 2020-07-22 2022-01-27 Edith Evert Device and Method for Imaging Structures Positioned Behind or Within a Surface
EP4099055A1 (en) * 2021-05-31 2022-12-07 Nokia Technologies Oy Synthetic aperture imaging
WO2023165678A1 (en) * 2022-03-01 2023-09-07 Telefonaktiebolaget Lm Ericsson (Publ) Method and apparatus for synthetic aperture radar scanning by a mobile communication device
WO2023247039A1 (en) * 2022-06-23 2023-12-28 Telefonaktiebolaget Lm Ericsson (Publ) Resource management of synthetic aperture radar in a mobile device

Family Cites Families (65)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US2735070A (en) 1956-02-14 riblet
CA1017001A (en) 1974-10-18 1977-09-06 Sed Systems Ltd. Fluid conductivity detecting means
US4211911A (en) 1979-01-16 1980-07-08 General Electric Company Microwave directional coupler and detector module
US4211224A (en) 1979-06-12 1980-07-08 Kubach John S Filter venting devices for ostomy appliances
US4626805A (en) 1985-04-26 1986-12-02 Tektronix, Inc. Surface mountable microwave IC package
JPH033289A (en) 1989-05-30 1991-01-09 Gurafuiko:Kk Twisted printed wiring
US5101163A (en) 1989-10-04 1992-03-31 Agar Corporation Ltd. Oil/water measurement
US5389735A (en) 1993-08-31 1995-02-14 Motorola, Inc. Vertically twisted-pair planar conductor line structure
US5572160A (en) 1994-12-01 1996-11-05 Teradyne, Inc. Architecture for RF signal automatic test equipment
US5774801A (en) 1995-08-23 1998-06-30 Ericsson Inc. High dynamic range mixer having low conversion loss, low local oscillator input power, and high dynamic range and a method for designing the same
US5829522A (en) 1996-07-18 1998-11-03 Halliburton Energy Services, Inc. Sand control screen having increased erosion and collapse resistance
US5718208A (en) 1996-09-16 1998-02-17 Ford Motor Company Fuel vapor management system
US6512475B1 (en) 1999-04-02 2003-01-28 Geophysical Survey Systems, Inc. High-frequency dual-channel ground-penetrating impulse antenna and method of using same for identifying plastic pipes and rebar in concrete
US7260369B2 (en) 2005-08-03 2007-08-21 Kamilo Feher Location finder, tracker, communication and remote control system
US6636816B1 (en) 1999-09-21 2003-10-21 Steven L. Dvorak Vector signal analysis method and apparatus therefor
JP3642512B2 (en) 2001-01-05 2005-04-27 アルケア株式会社 How to get mold material
US6784671B2 (en) 2002-02-04 2004-08-31 Mississippi State University Moisture and density detector (MDD)
US7164105B2 (en) 2002-04-05 2007-01-16 Microwave Imaging Systems Technologies, Inc. Non-invasive microwave analysis systems
US20040077943A1 (en) 2002-04-05 2004-04-22 Meaney Paul M. Systems and methods for 3-D data acquisition for microwave imaging
US7508421B2 (en) 2002-06-24 2009-03-24 Fujifilm Corporation Image pickup apparatus and image processing method
US8095204B2 (en) 2002-08-09 2012-01-10 Interstitial, Llc Apparatus and method for diagnosing breast cancer including examination table
US20040190377A1 (en) 2003-03-06 2004-09-30 Lewandowski Robert Stephen Method and means for isolating elements of a sensor array
US6806821B2 (en) * 2003-03-12 2004-10-19 Itt Manufacturing Enterprises, Inc. Apparatus and method for rapid detection of objects with time domain impulsive signals
US7034548B2 (en) 2003-04-11 2006-04-25 Agilent Technologies, Inc. Balanced device characterization including test system calibration
US7040139B2 (en) 2003-06-10 2006-05-09 Smiths Detection Inc. Sensor arrangement
US7119553B2 (en) 2003-06-11 2006-10-10 Konsulteurope Limited Limited Joint Stock Company Security scanners with capacitance and magnetic sensor arrays
US7676194B2 (en) * 2003-08-22 2010-03-09 Rappaport Theodore S Broadband repeater with security for ultrawideband technologies
US7527843B2 (en) 2004-04-28 2009-05-05 L'oreal Support provided with two series of labels
US7307575B2 (en) * 2004-09-14 2007-12-11 Bae Systems Information And Electronic Systems Integration Inc. Through-the-wall frequency stepped imaging system utilizing near field multiple antenna positions, clutter rejection and corrections for frequency dependent wall effects
US8050740B2 (en) 2004-09-15 2011-11-01 Wisconsin Alumni Research Foundation Microwave-based examination using hypothesis testing
US7148702B2 (en) 2005-03-30 2006-12-12 Agilent Technologies, Inc. VNA and method for addressing transmission line effects in VNA measurement data
US20060220658A1 (en) 2005-03-31 2006-10-05 Seichi Okamura Sensor system for moisture and salt measurement using microstripline
US7298312B2 (en) * 2005-06-27 2007-11-20 Itt Manufacturing Enterprises, Inc. Detecting small, time domain impulsive communications signals
US7755360B1 (en) * 2005-10-24 2010-07-13 Seektech, Inc. Portable locator system with jamming reduction
JP4834385B2 (en) 2005-11-22 2011-12-14 株式会社日立製作所 Printed circuit board and electronic device
AU2007207324B2 (en) * 2006-01-17 2011-05-26 Teledyne Uk Limited Surveillance apparatus and method
US7920088B2 (en) * 2006-03-03 2011-04-05 Scott Randall Thompson Apparatus and method to identify targets through opaque barriers
JP4898472B2 (en) 2006-04-11 2012-03-14 キヤノン株式会社 Inspection device
JP2008096112A (en) * 2006-10-05 2008-04-24 Denso Corp Radar system
US8302802B2 (en) 2007-05-31 2012-11-06 Honda Motor Co. Ltd. Installation of gas absorbent in a high pressure tank
US7668046B2 (en) 2007-06-06 2010-02-23 Christian Banker Apparatus and method for determining the position of an object in 3-dimensional space
US7852257B2 (en) * 2007-12-20 2010-12-14 Honeywell International Inc. Anti-interference microwave motion sensor
US7872604B2 (en) * 2007-12-20 2011-01-18 Honeywell International Inc. System and method for reducing interference in microwave motion sensors
US8260319B2 (en) * 2008-06-10 2012-09-04 Rosemount Tank Radar Ab Wireless transceiver
US9002427B2 (en) 2009-03-30 2015-04-07 Lifewave Biomedical, Inc. Apparatus and method for continuous noninvasive measurement of respiratory function and events
EP2431752A4 (en) 2009-05-11 2014-05-14 Panasonic Corp Antenna evaluation device and method
US8432322B2 (en) 2009-07-17 2013-04-30 Apple Inc. Electronic devices with capacitive proximity sensors for proximity-based radio-frequency power control
EP2480878B1 (en) 2009-09-22 2019-04-17 Adem Impedance sensing systems and methods for use in measuring constituents in solid and fluid objects
US9229102B1 (en) * 2009-12-18 2016-01-05 L-3 Communications Security And Detection Systems, Inc. Detection of movable objects
WO2011075639A1 (en) * 2009-12-18 2011-06-23 Christopher Gary Sentelle Moving entity detection
BR112012024337A2 (en) 2010-03-26 2017-10-03 Vayyar Imaging Ltd METHOD AND APPARATUS FOR ENHANCEMENT OF MICROWAVE IMAGE FORMATION OF AN OBJECT, MIMO RADAR SYSTEM, AND COMPUTER PROGRAM PRODUCT
US20130141287A1 (en) 2010-04-22 2013-06-06 Nokia Siemens Networks Oy Apparatus for Measuring a Radiation Pattern of an Active Antenna Arrangement
US20130300573A1 (en) * 2010-05-20 2013-11-14 Lifeflow Technologies, Inc. Patient monitoring and surveillance system, methods, and devices
US8620238B2 (en) 2010-07-23 2013-12-31 Blackberry Limited Method of power amplifier switching power control using post power amplifier power detection
US8917199B2 (en) * 2011-04-13 2014-12-23 Raytheon Company Subterranean image generating device and associated method
US9121943B2 (en) * 2011-05-23 2015-09-01 Sony Corporation Beam forming device and method
US9633247B2 (en) 2012-03-01 2017-04-25 Apple Inc. Electronic device with shared near field communications and sensor structures
US20130271328A1 (en) 2012-04-16 2013-10-17 Joshua G. Nickel Impedance Reference Structures for Radio-Frequency Test Systems
US9615765B2 (en) 2012-09-04 2017-04-11 Vayyar Imaging Ltd. Wideband radar with heterogeneous antenna arrays
US9223018B2 (en) * 2012-12-19 2015-12-29 Sony Corporation Method for displaying an active radar image and handheld screening device
US9084124B2 (en) 2012-12-21 2015-07-14 Apple Inc. Methods and apparatus for performing passive antenna testing with active antenna tuning device control
US9998684B2 (en) 2013-08-16 2018-06-12 Indiana University Research And Technology Corporation Method and apparatus for virtual 3D model generation and navigation using opportunistically captured images
US9261592B2 (en) * 2014-01-13 2016-02-16 Mitsubishi Electric Research Laboratories, Inc. Method and system for through-the-wall imaging using compressive sensing and MIMO antenna arrays
US10288728B2 (en) * 2015-04-29 2019-05-14 Vayyar Imaging Ltd System, device and methods for localization and orientation of a radio frequency antenna array
US10218052B2 (en) 2015-05-12 2019-02-26 Apple Inc. Electronic device with tunable hybrid antennas

Also Published As

Publication number Publication date
US20190260134A1 (en) 2019-08-22
US10854990B2 (en) 2020-12-01
US10566699B2 (en) 2020-02-18
US10290948B2 (en) 2019-05-14
US20170033469A1 (en) 2017-02-02
US20200161772A1 (en) 2020-05-21

Similar Documents

Publication Publication Date Title
US10854990B2 (en) System and method for radio frequency penetration imaging of an object
US11067685B2 (en) System, device and methods for imaging of objects using electromagnetic array
US11860262B2 (en) System and methods for three dimensional modeling of an object using a radio frequency device
US11709255B2 (en) System, device and methods for localization and orientation of a radio frequency antenna array
US20190025403A1 (en) System and methods for calibrating an antenna array using targets
US11480535B2 (en) System, device and methods for measuring substances′ dielectric properties using microwave sensors
JP2017528621A (en) Handheld multi-sensor system for measuring irregular objects
US9897699B2 (en) Methods and apparatus for virtual sensor array
US8957810B2 (en) Near-field millimeter wave imaging
US20230132993A1 (en) Radar based sensing for retail applications
US20160356877A1 (en) System and method for measurung an object location and/or orientation using radio frequency propagation measurements
US11580725B2 (en) Systems, devices and methods for imaging objects within or behind a medium using electromagnetic array
US20200166626A1 (en) Information processing device, information processing method, and storage medium having program stored thereon
Ortman et al. Real-time, mixed-mode computing architecture for waveform-resolved lidar systems with total propagated uncertainty
CN117930153A (en) Simulation method for detecting asteroid by circular orbit surrounding radar
JP2019124489A (en) Radar reflection cross-sectional area calculation apparatus, radar reflection cross-sectional area calculation method, and radar reflection cross-sectional area calculation program
JP2011191097A (en) Device, method, and program of specifying dielectric constant

Legal Events

Date Code Title Description
AS Assignment

Owner name: VAYYAR IMAGING LTD, ISRAEL

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ROSENFELD, JONATHAN;REEL/FRAME:054400/0789

Effective date: 20130109

Owner name: VAYYAR IMAGING LTD, ISRAEL

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:HOFFMAN, DAMIAN;REEL/FRAME:054399/0729

Effective date: 20181104

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION