CN107153247A - The vision sensing equipment of unmanned machine and the unmanned machine with it - Google Patents

The vision sensing equipment of unmanned machine and the unmanned machine with it Download PDF

Info

Publication number
CN107153247A
CN107153247A CN201710538808.9A CN201710538808A CN107153247A CN 107153247 A CN107153247 A CN 107153247A CN 201710538808 A CN201710538808 A CN 201710538808A CN 107153247 A CN107153247 A CN 107153247A
Authority
CN
China
Prior art keywords
image acquiring
unmanned machine
supporting part
sensing equipment
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
CN201710538808.9A
Other languages
Chinese (zh)
Inventor
刘少山
张哲�
李忠伟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Pusi Yingcha Technology Co Ltd
Original Assignee
Shenzhen Pusi Yingcha Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Pusi Yingcha Technology Co Ltd filed Critical Shenzhen Pusi Yingcha Technology Co Ltd
Priority to CN201710538808.9A priority Critical patent/CN107153247A/en
Publication of CN107153247A publication Critical patent/CN107153247A/en
Withdrawn legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B7/00Mountings, adjusting means, or light-tight connections, for optical elements
    • G02B7/02Mountings, adjusting means, or light-tight connections, for optical elements for lenses
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C11/00Photogrammetry or videogrammetry, e.g. stereogrammetry; Photographic surveying
    • G01C11/04Interpretation of pictures
    • G01C11/30Interpretation of pictures by triangulation
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B13/00Optical objectives specially designed for the purposes specified below
    • G02B13/06Panoramic objectives; So-called "sky lenses" including panoramic objectives having reflecting surfaces

Landscapes

  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Optics & Photonics (AREA)
  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Remote Sensing (AREA)
  • Length Measuring Devices By Optical Means (AREA)

Abstract

This application discloses a kind of vision sensing equipment of unmanned machine and the unmanned machine with it, the vision sensing equipment of the unmanned machine includes:Supporting part;It is arranged at four image acquiring devices on supporting part, four image acquiring devices are respectively arranged at the both sides of supporting part, wherein, first image acquiring device and the second image acquiring device are arranged at the first side of the supporting part, and the 3rd image acquiring device and the 4th image acquiring device are arranged at the second side of the supporting part;Control unit, described control unit is connected with four image acquiring devices respectively, described control unit is used to receive the image that four image acquiring devices are obtained, and the distance between target location and vision sensing equipment are determined with the image obtained according to four image acquiring devices.Thereby, it is possible to which to carrying out wide-angle deep vision measurement, apparatus structure is simple, and cost is low.

Description

The vision sensing equipment of unmanned machine and the unmanned machine with it
Technical field
The application is related to unmanned technology field, and more particularly to a kind of vision sensing equipment of unmanned machine and one kind have it Unmanned machine.
Background technology
With the rise in robot and unmanned field, allow robot and automobile perceive around the world into For academic research and the much-talked-about topic of enterprise innovation.Wherein, the purpose of perception is to allow the equipment such as robot and unmanned vehicle to know The location of oneself, understands the environment of surrounding, and the foundation of decision-making is provided for ensuing take action of equipment.In correlation technique, machine The main sensing solutions in device people and unmanned field are laser radar and the major class of vision two.Laser radar is using laser as light Source, remote sensing survey is completed by exploring laser light with the detected lightwave signal without interaction, can be used to map making, be determined Position and avoidance;Vision is camera, is widely used in the scenes such as object identification and object tracking.
But, the problem of correlation technique is present:One is, laser radar yields is low, and manufacturing cost is high, and big in order to measure Angular range, the mechanical part of rotation need to be set in laser radar causes service life to be limited;Two are, use monocular or binocular Camera can not realize that panoramic picture is obtained, and multi-cam is circular layout and places and can not unify because of different terminals, causes camera Between calibration amount and difficulty it is huge.
Therefore, the vision measurement technology of unmanned machine needs to improve.
The content of the invention
The application is intended at least solve one of technical problem in correlation technique to a certain extent.Therefore, the application One purpose is to propose a kind of vision sensing equipment of unmanned machine and the unmanned machine with it, being capable of vision on a large scale Measurement, simple in construction, cost is low.
Further object is to propose a kind of unmanned machine.
To reach above-mentioned purpose, the application one side embodiment proposes a kind of vision sensing equipment of unmanned machine, bag Include:Supporting part;Four image acquiring devices on supporting part are arranged at, four image acquiring devices are respectively arranged at support The both sides in portion, wherein, the first image acquiring device and the second image acquiring device are arranged at the first side of the supporting part, the 3rd Image acquiring device and the 4th image acquiring device are arranged at the second side of the supporting part;Control unit, described control unit It is connected respectively with four image acquiring devices, described control unit is used to receive the figure that four image acquiring devices are obtained Picture, the distance between target location and vision sensing equipment are determined with the image obtained according to four image acquiring devices.
According to the vision sensing equipment of the unmanned machine of the proposition of the embodiment of the present application, by by four image acquiring devices The both sides of supporting part are separately positioned on, and the image that four image acquiring devices are obtained is received by control unit, with according to four The image that individual image acquiring device is obtained determines the distance between target location and vision sensing equipment.Thus, the application is implemented Example can be to carrying out wide-angle deep vision measurement, and apparatus structure is simple, and cost is low.
According to one embodiment of the application, the first side of the supporting part can be the front side of supporting part, the of supporting part Two sides can be the rear side of supporting part.
According to one embodiment of the application, described first image acquisition device and second image acquiring device difference It is arranged at first end and the second end of the first side of the supporting part;And the 3rd image acquiring device and the 4th figure As acquisition device is respectively arranged at first end and the second end of the second side of the supporting part.
According to one embodiment of the application, the visible angle of the horizontal direction of the camera lens of four image acquiring devices More than 180 degree.
According to one embodiment of the application, four image acquiring devices can be fish eye lens.
According to one embodiment of the application, the supporting part can be cuboid or cylinder.
According to one embodiment of the application, the image obtained according to four image acquiring devices determines target The mode of the distance of position can be triangulation calculation.
To reach above-mentioned purpose, the application another aspect embodiment proposes a kind of unmanned machine, including affiliated nobody The vision sensing equipment of equipment.
The unmanned machine proposed according to the embodiment of the present application, can be to carrying out by the vision sensing equipment of unmanned machine Wide-angle deep vision is measured, and apparatus structure is simple, and cost is low.
According to one embodiment of the application, the vision sensing equipment of the unmanned machine is disposed vertically or horizontal positioned.
According to one embodiment of the application, the unmanned machine is robot or automatic driving vehicle.
Brief description of the drawings
Fig. 1 is the block diagram of the vision sensing equipment of the unmanned machine according to the embodiment of the present application;
Fig. 2 is the front view of the vision sensing equipment of the unmanned machine according to the application one embodiment;
Fig. 3 is the top view of the vision sensing equipment of the unmanned machine according to the application one embodiment;
Fig. 4 is the side view of the vision sensing equipment of the unmanned machine according to the application one embodiment;
Fig. 5 is the front view of the vision sensing equipment of the unmanned machine according to the application another embodiment;
Fig. 6 is the top view of the vision sensing equipment of the unmanned machine according to the application another embodiment;
Fig. 7 is the block diagram of the unmanned machine according to the embodiment of the present application.
Embodiment
Embodiments herein is described below in detail, the example of the embodiment is shown in the drawings, wherein from beginning to end Same or similar label represents same or similar element or the element with same or like function.Below with reference to attached The embodiment of figure description is exemplary, it is intended to for explaining the application, and it is not intended that limitation to the application.
The vision sensing equipment and unmanned machine of the unmanned machine of the embodiment of the present application are described with reference to the accompanying drawings.
Fig. 1 is the block diagram of the vision sensing equipment of the unmanned machine according to the embodiment of the present application.
As shown in figure 1, the vision sensing equipment of the unmanned machine of the embodiment of the present application includes:Supporting part 10, four images Acquisition device (i.e. the first image acquiring device 21, the second image acquiring device 22, the 3rd image acquiring device 23 and the 4th image Acquisition device 24) and control unit 30.
Wherein, four image acquiring devices are arranged on supporting part 10, and four image acquiring devices are respectively arranged at support The both sides in portion 10, the first image acquiring device 21 and the second image acquiring device 22 are arranged at the first side of supporting part 10, the 3rd The image acquiring device 24 of image acquiring device 23 and the 4th is arranged at the second side of supporting part 10, control unit 30 respectively with four Image acquiring device is connected, and control unit 30 is used to receive the image that four image acquiring devices are obtained, with according to four images The image that acquisition device is obtained determines the distance between target location and vision sensing equipment.
It should be noted that setting the first image acquiring device 21 and the second image to obtain dress in the first side of supporting part 10 22 are put, i.e., vision survey is carried out to the first side of supporting part 10 by the first image acquiring device 21 and the second image acquiring device 22 Amount, sets the 3rd image acquiring device 23 and the 4th image acquiring device 24 in the second side of supporting part 10, that is, passes through the 3rd figure As the image acquiring device 24 of acquisition device 23 and the 4th carries out vision measurement to the second side of supporting part 10.That is, support First side and the second side in portion 10 can carry out vision measurement by two vision sensing equipments respectively.
Specifically, first image acquiring device 21 and the second image acquiring device are set in the first side of supporting part 10 22, the second side of supporting part 10 sets the 3rd image acquiring device 23 and the 4th image acquiring device 24, and control unit 30 is received The image that four image acquiring devices are obtained, with the image obtained according to four image acquiring devices determine target location away from From.
Thus, the vision sensing equipment of unmanned machine can carry out wide-angle deep vision measurement, dress to supporting part both sides Put simple in construction, cost is low.Solving single camera can not realize that the measurement of Object Depth and dual camera can only obtain list The problem of image in one direction, while saving optional mechanical part built in laser radar, effectively save cost.
According to one embodiment of the application, as shown in figs. 4 and 6, the first side of supporting part 10 is the front side of supporting part 10, Second side of supporting part 10 is the rear side of supporting part 10.
It should be noted that before the front side of supporting part 10 and the rear side of supporting part 10 are respectively unmanned machine direct of travel Side and rear side, to ensure the vision measurement of the main traveling process of unmanned machine (moving forward and backward).That is, unmanned machine The setting direction of vision sensing equipment can be consistent with the direct of travel of unmanned machine, for example, when unmanned machine is robot, machine The direction of device people front institute direction can be the front side of supporting part 10, and the direction of back side institute of robot direction can be for after supporting part 10 Side, similarly, when unmanned machine is automatic driving vehicle, the direction of headstock institute direction can be the front side of supporting part 10, the tailstock The direction of institute's direction is the rear side of supporting part 10.So as to by respectively setting two images to obtain in the front side of supporting part 10 and rear side Take device to carry out visual sensing, it is ensured that the scope of visual sensing, improve the precision of visual sensing.
According to one embodiment of the application, the first image acquiring device 21 and the second image acquiring device 22 are set respectively First end and the second end in the first side of supporting part 10;And the 3rd image acquiring device 23 and the 4th image acquiring device 24 It is respectively arranged at first end and the second end of the second side of supporting part.
It should be noted that the first end of supporting part 10 and the distance between the second end of supporting part 10 can be according to applied fields Scape is adjusted, so that the visual range for being arranged at the image acquiring device at two ends ensures measured zone and overlapped as far as possible.
According to one embodiment of the application, the visible angle of the horizontal direction of the camera lens of four image acquiring devices is more than 180 degree.
Specifically, according to one embodiment of the application, four image acquiring devices are fish eye lens.
Specifically, when vision sensing equipment is disposed vertically (i.e. shown in Fig. 2), as shown in figure 3, the second image obtains dress Visual zone C, visual zone A and visual zone D visual pattern can be obtained by putting 22, and the 4th image acquiring device 24, which can be obtained, to be regarded Feel region C, visual zone B and visual zone D visual pattern, i.e. the second image acquiring device 22 and the 4th image acquiring device 24 can obtain visual zone C and visual zone D visual pattern.
When vision sensing equipment horizontal positioned (shown in Fig. 5), as shown in fig. 6, the first image acquiring device 21 and Two image acquiring devices 22 can obtain visual zone C, visual zone A and visual zone D visual pattern, and the 3rd image obtains dress Put 23 and the 4th image acquiring device 24 can obtain visual zone C, visual zone B and visual zone D visual pattern, i.e., first The image acquiring device 23 of image acquiring device 21 and the 3rd can obtain visual zone C visual pattern, the second image acquiring device 22 and the 4th image acquiring device 24 can obtain visual zone D visual pattern.
That is, 180 degree is more than by the visible angle of the horizontal direction of the camera lens of four image acquiring devices of setting, The vision measurement scope of vision sensing equipment can be made to be expanded to 360 degree of scopes.Wherein, visual zone E and visual zone F is four The blind spot of image acquiring device, because the region belongs to vision sensing equipment side and closer to the distance, therefore visual zone E and is regarded Feel that region F does not influence vision sensing equipment to carry out vision measurement.Wherein, visual zone E and visual zone F is obtained according to image and filled The visible angle of the horizontal direction for the camera lens put is determined.
According to one embodiment of the application, supporting part 10 is cuboid or cylinder, so that four image acquiring devices In two adjacent images acquisition device there is identical visual zone, and then vision measurement is carried out to visual zone.
According to one embodiment of the application, the image obtained according to four image acquiring devices determine target location away from From mode be triangulation calculation.
It should be noted that as shown in figs. 4 and 6, tested point X can carry out image acquisition by the first image acquiring device 21, And can be obtained by the second image acquiring device 22, i.e. the first image acquiring device 21 and the second image acquiring device 22 can be obtained simultaneously Tested point X image is taken, the distance of tested point X and vision sensing equipment is calculated by triangulation calculation (triangulation).Together Reason, tested point X, which may be disposed at two adjacent images acquisition device, can obtain the region of image jointly, that is to say, that four figures As acquisition device can be measured to any point in visual zone A, visual zone B, visual zone C and visual zone D, with basis The image that four image acquiring devices are obtained determines the distance between tested point X and vision sensing equipment.Wherein, tested point X can For the target location of vision measurement.
It should be appreciated that tested point X can be multiple, i.e., the image that control unit 20 is obtained according to image acquiring device is true The distance between fixed multiple tested point X and vision sensing equipment.Control unit 20 can be additionally used according to multiple tested point X and vision The distance between sensing device realizes VO (Visual odometry, visual odometry is calculated) algorithms and SLAM (Simultaneous Localization and Mapping, synchronously build figure and positioning) algorithm carries out deep learning to recognize that vision sensing equipment is regarded Feel the object in region.
In summary, according to the vision sensing equipment of the unmanned machine of the proposition of the embodiment of the present application, by the way that four are schemed As acquisition device is separately positioned on the both sides of supporting part, and the figure obtained by control unit four image acquiring devices of receiving Picture, the distance between target location and vision sensing equipment are determined with the image obtained according to four image acquiring devices.Thus, The embodiment of the present application can be to carrying out wide-angle deep vision measurement, and apparatus structure is simple, and cost is low.
The embodiment of the present application also proposed a kind of unmanned machine.
Fig. 7 is the block diagram of the unmanned machine according to the embodiment of the present application.
As shown in fig. 7, unmanned machine 200 includes the vision sensing equipment 100. of above-mentioned unmanned machine
According to one embodiment of the application, the vision sensing equipment 100 of unmanned machine is disposed vertically or horizontal positioned, i.e., The vision sensing equipment 100 of unmanned machine can be disposed vertically in unmanned machine, can be also placed horizontally in unmanned machine, but not It is only limitted to this.
According to one embodiment of the application, unmanned machine 200 can be robot or automatic driving vehicle.
The unmanned machine proposed according to the embodiment of the present application, can be to carrying out by the vision sensing equipment of unmanned machine Wide-angle deep vision is measured, and apparatus structure is simple, and cost is low.
In the description of the present application, it is to be understood that term " " center ", " longitudinal direction ", " transverse direction ", " length ", " width ", " thickness ", " on ", " under ", "front", "rear", "left", "right", " vertical ", " level ", " top ", " bottom " " interior ", " outer ", " up time The orientation or position relationship of the instruction such as pin ", " counterclockwise ", " axial direction ", " radial direction ", " circumference " be based on orientation shown in the drawings or Position relationship, is for only for ease of description the application and simplifies description, rather than indicate or imply that the device or element of meaning must There must be specific orientation, with specific azimuth configuration and operation, therefore it is not intended that limitation to the application.
In addition, term " first ", " second " are only used for describing purpose, and it is not intended that indicating or implying relative importance Or the implicit quantity for indicating indicated technical characteristic.Thus, define " first ", the feature of " second " can express or Implicitly include at least one this feature.In the description of the present application, " multiple " are meant that at least two, such as two, three It is individual etc., unless otherwise specifically defined.
In this application, unless otherwise clearly defined and limited, term " installation ", " connected ", " connection ", " fixation " etc. Term should be interpreted broadly, for example, it may be fixedly connected or be detachably connected, or integrally;Can be that machinery connects Connect or electrically connect;Can be joined directly together, can also be indirectly connected to by intermediary, can be in two elements The connection in portion or the interaction relationship of two elements, unless otherwise clear and definite restriction.For one of ordinary skill in the art For, the concrete meaning of above-mentioned term in this application can be understood as the case may be.
In this application, unless otherwise clearly defined and limited, fisrt feature can be with "above" or "below" second feature It is that the first and second features are directly contacted, or the first and second features pass through intermediary mediate contact.Moreover, fisrt feature exists Second feature " on ", " top " and " above " but fisrt feature are directly over second feature or oblique upper, or be merely representative of Fisrt feature level height is higher than second feature.Fisrt feature second feature " under ", " lower section " and " below " can be One feature is immediately below second feature or obliquely downward, or is merely representative of fisrt feature level height less than second feature.
In the description of this specification, reference term " one embodiment ", " some embodiments ", " example ", " specifically show The description of example " or " some examples " etc. means to combine specific features, structure, material or the spy that the embodiment or example are described Point is contained at least one embodiment of the application or example.In this manual, to the schematic representation of above-mentioned term not Identical embodiment or example must be directed to.Moreover, specific features, structure, material or the feature of description can be with office Combined in an appropriate manner in one or more embodiments or example.In addition, in the case of not conflicting, the skill of this area Art personnel can be tied the not be the same as Example or the feature of example and non-be the same as Example or example described in this specification Close and combine.
Although embodiments herein has been shown and described above, it is to be understood that above-described embodiment is example Property, it is impossible to the limitation to the application is interpreted as, one of ordinary skill in the art within the scope of application can be to above-mentioned Embodiment is changed, changed, replacing and modification.

Claims (10)

1. a kind of vision sensing equipment of unmanned machine, it is characterised in that including:
Supporting part;
Four image acquiring devices on supporting part are arranged at, four image acquiring devices are respectively arranged at the two of supporting part Side, wherein, the first image acquiring device and the second image acquiring device are arranged at the first side of the supporting part, and the 3rd image is obtained Device and the 4th image acquiring device is taken to be arranged at the second side of the supporting part;
Control unit, described control unit is connected with four image acquiring devices respectively, and described control unit is used to receive The image that four image acquiring devices are obtained, with the image obtained according to four image acquiring devices determine target location and The distance between vision sensing equipment.
2. the vision sensing equipment of unmanned machine according to claim 1, it is characterised in that the first side of the supporting part is The front side of supporting part, the second side of the supporting part is the rear side of supporting part.
3. the vision sensing equipment of unmanned machine according to claim 1, it is characterised in that described first image acquisition device First end and the second end of the first side of the supporting part are respectively arranged at second image acquiring device;And
3rd image acquiring device and the 4th image acquiring device are respectively arranged at the second side of the supporting part First end and the second end.
4. the vision sensing equipment of unmanned machine according to claim 1 or 3, it is characterised in that four images are obtained The visible angle of the horizontal direction of the camera lens of device is more than 180 degree.
5. the vision sensing equipment of unmanned machine according to claim 4, it is characterised in that four image acquiring devices For fish eye lens.
6. the vision sensing equipment of unmanned machine according to claim 1, it is characterised in that the supporting part be cuboid or Cylinder.
7. vision sensing equipment according to claim 1, it is characterised in that described according to four image acquiring devices The image of acquisition determines that the mode of the distance of target location is triangulation calculation.
8. a kind of unmanned machine, includes the vision sensing equipment of the unmanned machine according to claim any one of 1-7.
9. unmanned machine according to claim 8, it is characterised in that the vision sensing equipment of the unmanned machine is disposed vertically Or horizontal positioned.
10. unmanned machine according to claim 9, it is characterised in that the unmanned machine is robot or automatic driving car .
CN201710538808.9A 2017-07-04 2017-07-04 The vision sensing equipment of unmanned machine and the unmanned machine with it Withdrawn CN107153247A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710538808.9A CN107153247A (en) 2017-07-04 2017-07-04 The vision sensing equipment of unmanned machine and the unmanned machine with it

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710538808.9A CN107153247A (en) 2017-07-04 2017-07-04 The vision sensing equipment of unmanned machine and the unmanned machine with it

Publications (1)

Publication Number Publication Date
CN107153247A true CN107153247A (en) 2017-09-12

Family

ID=59796531

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710538808.9A Withdrawn CN107153247A (en) 2017-07-04 2017-07-04 The vision sensing equipment of unmanned machine and the unmanned machine with it

Country Status (1)

Country Link
CN (1) CN107153247A (en)

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10162362B2 (en) 2016-08-29 2018-12-25 PerceptIn, Inc. Fault tolerance to provide robust tracking for autonomous positional awareness
US10192113B1 (en) 2017-07-05 2019-01-29 PerceptIn, Inc. Quadocular sensor design in autonomous platforms
US10354396B1 (en) 2016-08-29 2019-07-16 Perceptln Shenzhen Limited Visual-inertial positional awareness for autonomous and non-autonomous device
US10366508B1 (en) 2016-08-29 2019-07-30 Perceptin Shenzhen Limited Visual-inertial positional awareness for autonomous and non-autonomous device
US10390003B1 (en) 2016-08-29 2019-08-20 Perceptln Shenzhen Limited Visual-inertial positional awareness for autonomous and non-autonomous device
US10395117B1 (en) 2016-08-29 2019-08-27 Trifo, Inc. Visual-inertial positional awareness for autonomous and non-autonomous tracking
US10402663B1 (en) 2016-08-29 2019-09-03 Trifo, Inc. Visual-inertial positional awareness for autonomous and non-autonomous mapping
US10410328B1 (en) 2016-08-29 2019-09-10 Perceptin Shenzhen Limited Visual-inertial positional awareness for autonomous and non-autonomous device
US10423832B1 (en) 2016-08-29 2019-09-24 Trifo, Inc. Visual-inertial positional awareness for autonomous and non-autonomous tracking
US10437252B1 (en) 2017-09-08 2019-10-08 Perceptln Shenzhen Limited High-precision multi-layer visual and semantic map for autonomous driving
US10453213B2 (en) 2016-08-29 2019-10-22 Trifo, Inc. Mapping optimization in autonomous and non-autonomous platforms
US10496104B1 (en) 2017-07-05 2019-12-03 Perceptin Shenzhen Limited Positional awareness with quadocular sensor in autonomous platforms
US10571926B1 (en) 2016-08-29 2020-02-25 Trifo, Inc. Autonomous platform guidance systems with auxiliary sensors and obstacle avoidance
US10571925B1 (en) 2016-08-29 2020-02-25 Trifo, Inc. Autonomous platform guidance systems with auxiliary sensors and task planning
TWI704408B (en) * 2018-08-22 2020-09-11 宏碁股份有限公司 Omnidirectional camera apparatus and image mapping/combining method thereof
US10794710B1 (en) 2017-09-08 2020-10-06 Perceptin Shenzhen Limited High-precision multi-layer visual and semantic map by autonomous units
US11314262B2 (en) 2016-08-29 2022-04-26 Trifo, Inc. Autonomous platform guidance systems with task planning and obstacle avoidance
US11774983B1 (en) 2019-01-02 2023-10-03 Trifo, Inc. Autonomous platform guidance systems with unknown environment mapping

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104079916A (en) * 2014-06-16 2014-10-01 深圳市德赛微电子技术有限公司 Panoramic three-dimensional visual sensor and using method
CN105314122A (en) * 2015-12-01 2016-02-10 浙江宇视科技有限公司 Unmanned aerial vehicle for emergency commanding and lane occupation evidence taking
CN106878687A (en) * 2017-04-12 2017-06-20 吉林大学 A kind of vehicle environment identifying system and omni-directional visual module based on multisensor
CN206946068U (en) * 2017-07-04 2018-01-30 深圳普思英察科技有限公司 The vision sensing equipment of unmanned machine and there is its unmanned machine

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104079916A (en) * 2014-06-16 2014-10-01 深圳市德赛微电子技术有限公司 Panoramic three-dimensional visual sensor and using method
CN105314122A (en) * 2015-12-01 2016-02-10 浙江宇视科技有限公司 Unmanned aerial vehicle for emergency commanding and lane occupation evidence taking
CN106878687A (en) * 2017-04-12 2017-06-20 吉林大学 A kind of vehicle environment identifying system and omni-directional visual module based on multisensor
CN206946068U (en) * 2017-07-04 2018-01-30 深圳普思英察科技有限公司 The vision sensing equipment of unmanned machine and there is its unmanned machine

Cited By (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11398096B2 (en) 2016-08-29 2022-07-26 Trifo, Inc. Visual-inertial positional awareness for autonomous and non-autonomous mapping
US11900536B2 (en) 2016-08-29 2024-02-13 Trifo, Inc. Visual-inertial positional awareness for autonomous and non-autonomous tracking
US10354396B1 (en) 2016-08-29 2019-07-16 Perceptln Shenzhen Limited Visual-inertial positional awareness for autonomous and non-autonomous device
US10366508B1 (en) 2016-08-29 2019-07-30 Perceptin Shenzhen Limited Visual-inertial positional awareness for autonomous and non-autonomous device
US10390003B1 (en) 2016-08-29 2019-08-20 Perceptln Shenzhen Limited Visual-inertial positional awareness for autonomous and non-autonomous device
US10395117B1 (en) 2016-08-29 2019-08-27 Trifo, Inc. Visual-inertial positional awareness for autonomous and non-autonomous tracking
US10571925B1 (en) 2016-08-29 2020-02-25 Trifo, Inc. Autonomous platform guidance systems with auxiliary sensors and task planning
US10410328B1 (en) 2016-08-29 2019-09-10 Perceptin Shenzhen Limited Visual-inertial positional awareness for autonomous and non-autonomous device
US10423832B1 (en) 2016-08-29 2019-09-24 Trifo, Inc. Visual-inertial positional awareness for autonomous and non-autonomous tracking
US11948369B2 (en) 2016-08-29 2024-04-02 Trifo, Inc. Visual-inertial positional awareness for autonomous and non-autonomous mapping
US10453213B2 (en) 2016-08-29 2019-10-22 Trifo, Inc. Mapping optimization in autonomous and non-autonomous platforms
US10769440B1 (en) 2016-08-29 2020-09-08 Trifo, Inc. Visual-inertial positional awareness for autonomous and non-autonomous tracking
US10496103B2 (en) 2016-08-29 2019-12-03 Trifo, Inc. Fault-tolerance to provide robust tracking for autonomous and non-autonomous positional awareness
US10571926B1 (en) 2016-08-29 2020-02-25 Trifo, Inc. Autonomous platform guidance systems with auxiliary sensors and obstacle avoidance
US11953910B2 (en) 2016-08-29 2024-04-09 Trifo, Inc. Autonomous platform guidance systems with task planning and obstacle avoidance
US11842500B2 (en) 2016-08-29 2023-12-12 Trifo, Inc. Fault-tolerance to provide robust tracking for autonomous and non-autonomous positional awareness
US10402663B1 (en) 2016-08-29 2019-09-03 Trifo, Inc. Visual-inertial positional awareness for autonomous and non-autonomous mapping
US11544867B2 (en) 2016-08-29 2023-01-03 Trifo, Inc. Mapping optimization in autonomous and non-autonomous platforms
US10832056B1 (en) 2016-08-29 2020-11-10 Trifo, Inc. Visual-inertial positional awareness for autonomous and non-autonomous tracking
US10929690B1 (en) 2016-08-29 2021-02-23 Trifo, Inc. Visual-inertial positional awareness for autonomous and non-autonomous mapping
US10943361B2 (en) 2016-08-29 2021-03-09 Trifo, Inc. Mapping optimization in autonomous and non-autonomous platforms
US10983527B2 (en) 2016-08-29 2021-04-20 Trifo, Inc. Fault-tolerance to provide robust tracking for autonomous and non-autonomous positional awareness
US11314262B2 (en) 2016-08-29 2022-04-26 Trifo, Inc. Autonomous platform guidance systems with task planning and obstacle avoidance
US11328158B2 (en) 2016-08-29 2022-05-10 Trifo, Inc. Visual-inertial positional awareness for autonomous and non-autonomous tracking
US10162362B2 (en) 2016-08-29 2018-12-25 PerceptIn, Inc. Fault tolerance to provide robust tracking for autonomous positional awareness
US11501527B2 (en) 2016-08-29 2022-11-15 Trifo, Inc. Visual-inertial positional awareness for autonomous and non-autonomous tracking
US10192113B1 (en) 2017-07-05 2019-01-29 PerceptIn, Inc. Quadocular sensor design in autonomous platforms
US10496104B1 (en) 2017-07-05 2019-12-03 Perceptin Shenzhen Limited Positional awareness with quadocular sensor in autonomous platforms
US10794710B1 (en) 2017-09-08 2020-10-06 Perceptin Shenzhen Limited High-precision multi-layer visual and semantic map by autonomous units
US10437252B1 (en) 2017-09-08 2019-10-08 Perceptln Shenzhen Limited High-precision multi-layer visual and semantic map for autonomous driving
TWI704408B (en) * 2018-08-22 2020-09-11 宏碁股份有限公司 Omnidirectional camera apparatus and image mapping/combining method thereof
US11774983B1 (en) 2019-01-02 2023-10-03 Trifo, Inc. Autonomous platform guidance systems with unknown environment mapping

Similar Documents

Publication Publication Date Title
CN206946068U (en) The vision sensing equipment of unmanned machine and there is its unmanned machine
CN107153247A (en) The vision sensing equipment of unmanned machine and the unmanned machine with it
CN105607635B (en) Automatic guided vehicle panoramic optical vision navigation control system and omnidirectional's automatic guided vehicle
US10748426B2 (en) Systems and methods for detection and presentation of occluded objects
WO2020258721A1 (en) Intelligent navigation method and system for cruiser motorcycle
CN106227212B (en) The controllable indoor navigation system of precision and method based on grating map and dynamic calibration
CN111201879A (en) Grain harvesting and transporting integrated loading device/method based on image recognition
US7652686B2 (en) Device for image detecting objects, people or similar in the area surrounding a vehicle
KR102034189B1 (en) Method and device for visualizing the surroundings of a vehicle
US20150219462A1 (en) Method and device for determining a vehicle position in a mapped environment
US20160195407A1 (en) Vehicle driving guidance device and method
CN109634279A (en) Object positioning method based on laser radar and monocular vision
WO2010004911A1 (en) Train-of-vehicle travel support device
CN106444837A (en) Obstacle avoiding method and obstacle avoiding system for unmanned aerial vehicle
CN108888187A (en) A kind of sweeping robot based on depth camera
CN109373975A (en) Detect vehicle control apparatus, control method and computer program
US20210270957A1 (en) Ranging method, ranging device and ranging system
US20150066364A1 (en) Navigation system
CN104217439A (en) Indoor visual positioning system and method
CN112698306A (en) System and method for solving map construction blind area by combining multiple laser radars and camera
CN106462314A (en) Dynamic camera view to aid with trailer attachment
CN110491156A (en) A kind of cognitive method, apparatus and system
CN109444916A (en) The unmanned travelable area determining device of one kind and method
US11774981B2 (en) Driver aid and autonomous tractor-trailer parking and loading dock alignment system
CN111353453B (en) Obstacle detection method and device for vehicle

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WW01 Invention patent application withdrawn after publication

Application publication date: 20170912

WW01 Invention patent application withdrawn after publication