US20220083070A1 - Identification And Localization Of A Base Station Of An Autonomous Mobile Robot - Google Patents

Identification And Localization Of A Base Station Of An Autonomous Mobile Robot Download PDF

Info

Publication number
US20220083070A1
US20220083070A1 US17/536,917 US202117536917A US2022083070A1 US 20220083070 A1 US20220083070 A1 US 20220083070A1 US 202117536917 A US202117536917 A US 202117536917A US 2022083070 A1 US2022083070 A1 US 2022083070A1
Authority
US
United States
Prior art keywords
base station
robot
navigation
map
user
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US17/536,917
Inventor
Reinhard Vogel
Harold Artes
Christoph Freudenthaler
Fabian Lenhardt
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Robart GmbH
Original Assignee
Robart GmbH
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Robart GmbH filed Critical Robart GmbH
Priority to US17/536,917 priority Critical patent/US20220083070A1/en
Assigned to RobArt GmbH reassignment RobArt GmbH ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ARTES, HAROLD, Freudenthaler, Christoph, LENHARDT, Fabian, VOGEL, REINHARD
Publication of US20220083070A1 publication Critical patent/US20220083070A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05DSYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
    • G05D1/00Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
    • G05D1/02Control of position or course in two dimensions
    • G05D1/021Control of position or course in two dimensions specially adapted to land vehicles
    • G05D1/0212Control of position or course in two dimensions specially adapted to land vehicles with means for defining a desired trajectory
    • G05D1/0225Control of position or course in two dimensions specially adapted to land vehicles with means for defining a desired trajectory involving docking at a fixed facility, e.g. base station or loading bay
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J9/00Programme-controlled manipulators
    • B25J9/0003Home robots, i.e. small robots for domestic use
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05DSYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
    • G05D1/00Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
    • G05D1/02Control of position or course in two dimensions
    • G05D1/021Control of position or course in two dimensions specially adapted to land vehicles
    • G05D1/0231Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means
    • G05D1/0234Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means using optical markers or beacons
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05DSYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
    • G05D1/00Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
    • G05D1/02Control of position or course in two dimensions
    • G05D1/021Control of position or course in two dimensions specially adapted to land vehicles
    • G05D1/0231Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means
    • G05D1/0238Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means using obstacle or wall sensors
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05DSYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
    • G05D1/00Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
    • G05D1/02Control of position or course in two dimensions
    • G05D1/021Control of position or course in two dimensions specially adapted to land vehicles
    • G05D1/0268Control of position or course in two dimensions specially adapted to land vehicles using internal positioning means
    • G05D1/0274Control of position or course in two dimensions specially adapted to land vehicles using internal positioning means using mapping information stored in a memory device
    • G05D2201/0215

Definitions

  • the application relates to a method for the identification and localization of a base station of an autonomous mobile robot by the robot. It is further described how the base station may be employed for the calibration of a sensor of the robot and how the docking maneuver may be improved.
  • aspect consists in simplifying or improving known methods for identifying and locating a robot base station by the robot as well as the known methods for docking onto the base station.
  • the robot comprises a navigation module with a navigation sensor for detecting the geometric features of objects in the environment of the robot.
  • the base station has at least one geometric feature that can be detected by the robot by means of the navigation sensor.
  • the robot includes a robot controller that is coupled to the navigation module and which is configured, based on the at least one geometric feature of the base station, to identify the base station and/or to localize and/or to determine a docking position of the robot.
  • a further example of the application relates to a base station for a mobile robot.
  • the base station has a housing with at least one opening arranged in the housing which, due to its geometry, defines at least one geometric feature that is detectable by the sensor system of the robot.
  • the method comprises the detection of geometric features of objects in the environment of the robot by means of a navigation module of the robot which has a navigation sensor. At least one of the thus detected objects is a geometric feature of the base station. The method further comprises the identification and/or localization of the base station based on the at least one geometric feature of the base station.
  • the robot comprises a navigation module with a navigation sensor for the detection of geometric features of objects in the environment of the robot.
  • the base station has at least one geometric feature that can be detected by the robot by means of a navigation sensor.
  • the navigation module is configured to test and/or calibrate the navigation sensor with the aid of the at least one detected geometric feature of the base station.
  • a method comprises the detection of geometric features in the environment of robot by means of a navigation module of the robot which has a navigation sensor, wherein at least one of the detected features is a geometric feature of the base station.
  • the navigation sensor is calibrated and/or tested with the aid of the at least one geometric feature of the base station.
  • a further method serves to dock an autonomous mobile robot onto a base station.
  • the method comprises the detection of a docking position of the robot on the base station, wherein the docking position comprises a location and an orientation of the robot, as well as the navigation of the robot into the docking position. Afterwards it is tested whether the robot has correctly docked onto the base station. If this is not the case, the position of the robot is varied and it is once again tested whether the robot has correctly docked onto the base station. The varying and testing is carried out until either the test is successful or an abort criterion is fulfilled.
  • a method for the automatic docking of an autonomous mobile robot onto a base station includes the detection of obstacles by means of a navigation module of the robot which has a navigation sensor, as well as testing whether, within a defined area around the base station, access to the base station for the robot is impeded by detected obstacles. If the test reveals that access to the base station is impeded for the robot, a disturbance is communicated via a user interface.
  • a further exemplary method for an autonomous mobile robot includes the detection of geometric features of objects in the environment of the robot by means of a navigation module of the robot which comprises a navigation sensor, as well as the navigation of the robot based on at least one of the detected geometric features and on an electronic map of the area in which the robot operates.
  • the position of a base station of the robot is designated on the electronic map.
  • the method further includes testing whether the detected geometric features contain a geometric feature that is linked with the base station. If this is the case, the current position of the base station is determined based on the geometric feature that is linked with the base station and the position of the base station is updated on the electronic map.
  • the base station may be detected and located in a different manner in order to update its position on the map of the robot.
  • a further example of a method for an autonomous mobile robot includes the detection of geometric features of objects in the environment of the robot by means of a navigation module of the robot which comprises a navigation sensor, as well as the navigation of the robot based on at least one of the detected geometric features and on an electronic map of the area in which the robot operates.
  • the position of a base station of the robot is designated on the electronic map.
  • a first geometric feature that is not defined by the base station is linked with the position of the base station.
  • This first geometric feature is tracked using a SLAM algorithm, wherein a position of the first geometric feature on the electronic map is kept current and the position of the base station is saved as a relative position relative to the position of the first geometric feature.
  • FIG. 1 shows a robot with a base station in its area of robot operation.
  • FIG. 2 schematically shows by means of a drawing the optical distance measurement using triangulation.
  • FIG. 3 shows an example of a base station with geometric features that are detectable by the navigation sensor of the robot and which are defined by openings in the front housing wall of the base station.
  • FIG. 4 illustrates the detection of the geometric figures of a base station using the navigation sensor of the robot, as well as possible systematic measurement errors.
  • FIG. 5 illustrates a method for docking a robot onto a base station, wherein the robot continues to vary its orientation until it has correctly docked.
  • a mobile robot should be able to dock onto its base station reliably and safely.
  • Known systems base stations and robots
  • methods for finding and recognizing a base station and its exact position and orientation often employ special sensors in the robot (e.g. a light beam), complex image processing algorithms, which must be executed in addition to the navigation algorithms, and/or special markings on the base station or in the area in which the robot is operated.
  • the locating of the base station and the docking of the robot onto the base station should be robust against many different kinds of disturbances. For example, shifting of the base station may impair a reliable localization of the base station.
  • the functionality of the sensors employed by the robot for navigation cannot be reliably tested in an unknown environment, such as in the area in which the robot is operated.
  • the approach to the base station (docking maneuver) may sometimes end in a failed charging contact due to odometry errors.
  • the user for example, may place obstacles near the base station that disrupt the docking maneuver and thus result in a failed docking.
  • the base station is recognized and localized with the aid of its geometric shape (e.g. certain geometric characteristics (outer form/shape) of the housing of the base station).
  • the base station itself exhibits, due to its outer form/shape, at least one geometric feature, with the aid of which the robot may recognize the base station.
  • Special markings on the housing of the base station (which nevertheless have no influence on its geometric form/shape) or the emission of a (locating) signal are no longer necessary.
  • a robot should quickly and reliably find its way back to its base station after completion of its (e.g. cleaning) task.
  • the position of the base station is marked on a map and is kept current by means of various methods.
  • a further desirable capability of the robot is that it be able to maintain its bearings (to accurately and quickly orient itself) in different areas in which the robot operates (such as, for example, on the various floors of a building).
  • one base station is employed in each of the areas in which the robot operates that the robot is able to distinguish from one another.
  • the robot links a map of the respective area of operation with each base station. Once the robot has identified a specific base station, it can immediately begin a (self) localization with the aid of the map that is linked with each base station.
  • a further desirable capability of a mobile robot is a robust and exact docking onto the base station so that, for example, a contact between the charging contacts of the robot and those of the base station can be reliably carried out.
  • the position and orientation of the robot is slightly corrected in order to be able to safely dock.
  • a docking should also be possible in the case of possible disturbance (e.g. access to the base station is blocked).
  • a possible disturbance is detected and the user is informed of it.
  • FIG. 1A shows an autonomous mobile robot 100 and a corresponding base station 200 in an operation area of the robot.
  • the robot has a navigation module with at least one navigation sensor 110 for orienting itself in its environment and for navigating throughout the operation area of the robot, thus enabling it to autonomously carry out a task.
  • the navigation module of the robot 100 collects with the aid of the navigation sensor 110 information regarding the position of “navigation features” in the environment of the robot, e.g. geometric features of objects such as, e.g. those of obstacles, as well as information regarding the floor covering, etc.
  • Geometric features are, for example, surfaces (e.g. a wall), lines (e.g.
  • Every navigation feature can be associated with a position (including orientation) in the room and, if needed, recorded on the electronic map of the robot 100 .
  • the navigation module operates, for example, using an obstacle avoidance strategy and/or a SLAM algorithm (Simultaneous Localization and Mapping) and/or one or more maps of the area in which the robot operates.
  • the robot can create a new map of the robot operation area while in operation or it can make use of an already existing map before beginning its task.
  • An already existing map can be created by the robot itself during the completion of an early task, such as an exploratory run, or may be provided by another robot and/or a person.
  • the navigation sensor 110 detects information about the environment of the robot (and thus about the area in which it operates), in particular about its geometric features in one, two or three dimensions.
  • the navigation sensor 110 may be, for example, a sensor for distance measurement such as, for example, an optical and/or acoustic sensor that detects the distance between the sensor and an obstacle by means of triangulation or by measuring the travel time of an emitted signal (e.g. a laser beam or an ultrasonic signal). Accordingly, e.g. triangulation sensors, time of flight cameras, laser scanners, ultrasonic sensors and the like may be employed as navigation sensors.
  • the navigation module of the robot 100 can determine the distance between the robot and the individual points, lines and/or surfaces of objects in the environment of the robot.
  • the thus detected points, lines and/or surfaces are designated as “navigation features” and are saved on a map of the robot (i.e. the geometry of the navigation feature and its position, including orientation, is designated on the electronic map). Later, the robot can orient itself with the aid of these navigation sensors and/or avoid the detected obstacles, thus preventing collision.
  • a navigation sensor 110 is a camera (e.g. with a CCD or CMOS image sensor) that has a detection module for identifying the corners and edges of object (navigation features) in the environment recorded on an image by means of image data processing.
  • the position of a navigation feature within a projected plane relative to the robot can be determined.
  • the position, including the orientation, of the navigation feature and that of the robot within the room can be determined on the basis of this movement.
  • This procedure is known as visual SLAM.
  • FIG. 1C shows a simple example of a base station 200 . It consists of a main body 210 with geometric features that can be detected by the navigation sensor as navigation features.
  • the main body 210 of the base station 200 may include various further components such as, for example, components for charging the battery of the robot 100 or for vacuuming out dirt.
  • Two charging contacts 220 are arranged on the base station 200 . Corresponding contact are arranged on the robot 100 which must come into contact with the charging contacts 220 in order to allow a successful charging of the robot battery. This means that, in order to charge its batter, the robot must dock onto the base station at a certain position and in a certain orientation. The position and orientation required for this (and, thus, the path to be followed), can be determined by the robot when the position and orientation of the main body 200 are known to it.
  • the geometric characteristics of the base station 200 are detected as navigation features with the aid of the navigation sensor 110 of the robot for the purpose of its navigation.
  • the navigation features of the base station can be definitively and without a great deal of additional effort identified, their position and orientation in the robot's environment can be determined and they can be recorded on the map of the robot.
  • Simple criteria are used to achieve this purpose such as, for example, (1.) the distance separating certain points (e.g.
  • corners of a feature from other features, (2.) the length of edges/segments/lines, (3.) the size of surfaces, (4.) the location of features relative to each other, in particular the angle between two edges, segments or lines or the angle defined by three specific points (corners), (5.) ratios (relative sizes, lengths), e.g.
  • error metrics such as the (squared) deviation of the (erroneous) measurement from the construction-related standard values (with the aid of a larger quantity of points, a straight line, for example, can be determined that can be established by means of linear regression; the degree of deviation of a specific point from the regression lines can serve as a criterion as to whether the given point altogether belongs to the assumed line).
  • the detected features designate a width, a depth and/or a height of a part of the main body of the base station.
  • the criteria used to locate the base station should be easy to calculate and be robustly determinable from various positions. This is explained in greater detail in the following.
  • the distance to obstacles in the environment of the robot is determined, e.g. by means of optic triangulation, for the purpose of which structured light (in a visible or non-visible range) in the form of a line running approximately parallel to the surface of the floor is emitted.
  • the principle of optical triangulation is illustrated in FIG. 2 .
  • the emitted structured light 111 strikes an obstacle H and is diffusely scattered off of it.
  • a camera takes a picture of the environment that depicts the light that is reflected back from the obstacle (e.g. a light line).
  • the distance d to the illuminated obstacles H at selected points or along the entire line can be determined based (in the case of a horizontal light line) on the vertical position x of the structured, reflected light (i.e. of the light line) in the picture.
  • the base station 200 is geometrically designed such that, because of its height, it is illuminated by the light emitted from the navigation sensor 110 and, because of its depth, is clearly distinguishable from a wall. Furthermore, the surface of the base station is designed such that the light emitted from the navigation sensor 110 is reflected in a well diffused state (meaning that it exhibits no absorbing or mirroring surfaces).
  • FIG. 1B shows, as an example, the possible results of a sensor measurement of the scenario shown in FIG. 1A .
  • the robot 100 recognizes with the aid of its navigation modules and its navigation sensors 110 two line shaped navigation FIGS. 201 and 202 stemming from the measurement of the base station.
  • the navigation features 201 , 202 will be designated as boundary lines or, in short, as lines.
  • These two lines 201 and 202 each have a characteristic length which corresponds to the width (in horizontal direction) of the side surface (line 201 ) and of the front surface (line 202 ).
  • the two lines 201 and 202 include a certain angle (e.g. a right angle). If, for example, the base station 200 (e.g.
  • the robot 100 in a horizontal plane that lies at a certain height above the floor) has a rectangular cross sectional area having side lengths of, e.g. 5 cm and 15 cm and the robot 100 , with the aid of its navigation sensor 110 , detects two almost rectangular lines having lengths of nearly 5 cm (line 201 ) and 15 cm (line 202 ), then the robot 100 can interpret these lines as the base station.
  • use may be made of the side ratio of 5:15 1:3.
  • one side of the base station cannot be recognized by the robot due to the shadowing caused by the front surface of the navigation sensor.
  • the robot can move round the base station until it has also reliably recognized the second side surface.
  • the base station will be regarded from at least two different positions.
  • the base station 200 is recognized solely based on its rectangular form. In such cases, the probability for an erroneous detection is high, as any rectangular object (for example, a box) having the corresponding side lengths would be recognized as the base station. Furthermore, the described corners will have generally been replaced by rounded edges, both for safety reasons (risk of injury), as well as for reasons of design. In principle, such rounded edges are also detectable with the aid of the navigation sensor described above. The calculation effort needed for the processing of rounded forms, however, is greater than that needed for straight lines.
  • the inside of the base station is also utilized for its detection.
  • one or more openings are introduced into the main body of the base station. Individual components (for example, for vacuuming dirt) of the base station may be visible through these openings. This may also be omitted, however, so that only one or more inner walls can be recognized.
  • FIG. 3A shows an example of a base station 200 with rounded edges and two openings 230 arranged in the front side of the base station.
  • FIG. 3B shows a cross section view through the base station 200 of FIG. 3A , wherein the section plane is a horizontal plane that lies at a certain height above the floor. This certain height is the height at which the navigation sensor 110 carries out the distance measurement to obstacles in the environment of the robot.
  • FIG. 3C shows an example of the results of such a measurement when the robot is standing at a certain distance (for example 0.5-1 m) in front of the base station. From here it can see parts of the back wall of the base station 200 , which all lie along one line.
  • the robot can determine the position of four points (1, 2, 3, 4) with negligible measurement errors.
  • the position and size of the two openings 230 has been chosen in this case such that the distance separating any of the two points is different for each pair of points, (meaning, in particular, that d(1; 2) ⁇ d(1; 3) ⁇ d(1; 4) ⁇ d(2; 3) ⁇ d(2; 4) ⁇ d(3; 4), wherein d(i, j) designates the distance between point i and point j).
  • These distances can be easily calculated and compared to stored standard values.
  • the distance separating the points from the line formed by the back wall may also be compared, for example.
  • the ratios of the distances to each other may also be taken into consideration.
  • an error function such as, for example, the coefficient of determination of a regression model, the degree of exactness with which the points 1, 2, 3, and 4 lie along a line parallel to the back wall can be tested.
  • a cover 250 made of a transparent material that allows the light emitted from the navigation sensor 110 to pass through, may be arranged over the openings 230 .
  • the emitted light may lie, for example, in an (infrared) spectrum that is not visible to humans, so that the cover as viewed by the robot is transparent, but, as viewed by humans, will appear colored and nontransparent.
  • Such a cover 250 should not only cover the openings 230 , but should also, independent of this, be used to form differing geometric shapes that can be recognized by the user and the robot. In this manner, aspects of design may be combined with a simplified detectability.
  • the cover 250 may be provided with an antireflective coating that is matched to the wavelength of the light emitted by the navigation sensor 110 .
  • the base station may be reliably recognized in this plane based on its geometric form, at least a selection of the recognizable navigation features (or more specifically, their underlying dimensions) must be found in the environment of the expected cross section measurement, irrespective of the actual height of measurement (the distance of the plane in which the distance measurement is carried out from the floor).
  • the robot will be employed in two mutually isolated areas of robot operation such as, for example, on two different floors of a building.
  • a base station 200 In each of the areas of operation a base station 200 is located that the robot 100 can definitively recognize with the aid of its navigation sensor 110 . If the robot 100 is thus now capable of differentiating between the base stations 200 , it also immediately obtains information as to which of the various areas of operation it is located in without the need for the user to additionally intervene.
  • some parts of the geometric form of the base station may be altered by the user (e.g. at the time the base station is put into operation), meaning that a feature that is recognizable by the navigation sensor 110 may be altered in a previously defined manner.
  • This alteration can be achieved, for example, by shifting, removing or adding parts to the base station.
  • an alteration can be carried out by changing the size (width) of one of the two openings, for example.
  • the location of point 3 ( FIG. 3C ) can be varied by adding, for example, an additional screen.
  • a sliding screen may be arranged behind the centre panel between the two viewing windows.
  • Autonomous mobile robots 100 that mark their base stations 200 on a map are basically known. Often the base station serves as the starting point of the map, i.e. the base station has a permanent position on the map. The position of the base station, however, can change. Detecting the base station based on characteristics (navigation features) that can be detected by the navigation sensor 110 makes it possible to easily update the position of the base station on the map. However, it is also possible to detect the base station by other means, thereby allowing the position of the base station on the map of the robot to be updated. The change in the position of the base station may be carried out by a user, e.g. (1.) while the robot is in operation or (2.) while the robot is docked on the base station.
  • a user e.g. (1.) while the robot is in operation or (2.) while the robot is docked on the base station.
  • mapping carried out by the robot and/or its localization on the map may lead to the result that the position of the base station anticipated by the robot does not correspond to its actual position. In the worst case, this can result in the loss of all (meaningful) information about the position of the base station.
  • examples will be used to explain how this problem can be solved, or at least mitigated.
  • Position updates using a SLAM algorithm SLAM algorithms (Simultaneous Localization and Mapping), by means of which the position of the robot and selected navigation features that have been detected by the navigation sensor are continuously correlated, are often employed for the navigation of autonomous mobile robots. This enables to robot to carry out a robust mapping despite measurement and odometry errors.
  • SLAM demands a relatively high amount of computing capacity, which is why only a few selected and readily visible navigation features such as, for example, walls are taken into consideration in order to limit the needed calculation effort.
  • the positions of the navigation features (and those of the robot) that are tracked in the SLAM algorithm are continuously corrected; meaning they are at no permanent (relative) position in relation to other objects that are not tracked in the SLAM algorithm.
  • the features of the base station are generally too small to be tracked by the SLAM algorithm and they may therefore shift their recorded position relative to a wall (which is tracked by the SLAM algorithm). This can result in the recorded position being located in the wrong room or even outside of the area of operation. For this reason it is useful to update the position of the base station.
  • the position of the base station can be easily kept current using the SLAM algorithm.
  • at least one easily detectable feature of the base station (cf. FIG. 3 ) is selected and tracked with the aid of the SLAM algorithm.
  • This can be, for example, the segment formed by points 2 and 3 in FIG. 3C .
  • a feature can also be used that has been detected in the proximity of the base station. This can be, for example, the line 300 in FIG. 1B , which is formed by the wall against which the base station stands.
  • the robot records which of the features tracked by the SLAM algorithm determines the position of the base station.
  • the robot can save the position of the base station relative to this navigation feature. When the robot wants to return to the base it can move to a position near this feature, thus ensuring that it will reliably find its way back again.
  • Position updates during operation of the robot In the event that the base station, for example, is moved by the user while the robot is in operation, it is advantageous for the robot 100 to recognize the base station 200 when moving past it and to update the map with the new position. Since the navigation sensor detects features of the base station to navigate the robot, it is sufficient in this case for the newly detected features to be tested as to whether they might be part of the base station. In this case, as a first step, an easily recognized feature, for the detection of which only a small amount of additional calculation time is required, can be used. After recognition has been confirmed, further characteristics of the base station can be tested.
  • segments e.g. lines of a certain length and orientation
  • segments can be looked for that correspond in their length to the distance separating points 2 and 3 in the FIG. 3C .
  • it can be further tested as to whether a back wall is recognizable, whether this is at the correct distance, and whether the points 1 and 4 are present and lie along one line. While doing so, the length of recognized segments, for example, can be routinely detected by the navigation module in order to test them regarding their relevance for the navigation.
  • the new position of the base station is recorded and the former one is deleted. This is particularly useful when the navigation sensor simultaneously examines the former position of the base station and no base station is detected at this position. If a base station is detected at a location in the area of robot operation that is far away from the previous base station, then this may be due to the presence of a second base station. In this case the position of the base station is newly recorded and the former position is retained for later verification. When the robot is in the proximity of the former position, it tests whether the base station is still present. Depending on the result, the former position is either deleted or it is recorded that there are two base stations located in the area of operation of the robot. The respective positions (including orientation) of both base stations can be saved on the map.
  • the robot finds itself in the proximity of a base station but does not detect it. This may be the case if the position of the base station has been changed by the user, but also if the user has only temporarily removed it, for example to clean it. In the simplest case the robot will delete the position of the base station from the map data. Alternatively, this position on the robot's map may be marked as “questionable”. If while carrying out its operational task the robot later recognizes a base station at a different position, then the position marked as questionable is deleted and the new position of the base station is recorded. In the opposing case, the robot returns to the position marked as questionable when it wants to return to the base station.
  • a special search program is started to find the base station, wherein the robot moves around the area of operation and looks specifically for features (navigation features) of the base station.
  • a message can be sent to the user (e.g. by means of a user interface, e.g. an application on a tablet computer or mobile telephone), allowing the user to react to the situation, if needed.
  • the measurement accuracy of the navigation sensor 110 may deteriorate as the distance to the object to which the distance measurement is to be carried out increases. In the case of object with elongated dimensions, the accuracy also depends on their orientation. This may particularly be the case when triangulation sensors are used. If, for example, the base station 200 of FIG. 3 is viewed directly from the front, then the distance separating the points 2 and 3 can be determined at a distance of approximately 2 m with sufficient accuracy. If the base station, however, is viewed from an angle of 45°, then, in order to ensure a reliable measurement, the distance between the base station 200 and the robot 100 should not exceed about 1 m.
  • a maximum distance for testing whether a detected navigation feature belongs to the base station can be determined in order to improve the quality of the detection of the base station.
  • the maximum distance for carrying out a test may also depend on the orientation of the feature (relative to the robot).
  • Position updates at the start of robot operation A user may move the robot together with the base station while the robot is docked onto it and while the robot, for example, is in pause modus. If the robot is using an existing map that, for example, it had compiled while carrying out a previous operational task, then its new position (and that of the base station) will no longer correspond with the existing navigation information.
  • the robot may carry out the following steps at the start of its operational task: (1.) Exit the base station and collect map data that is relevant for its operational task; (2.) Localization of the robot in the existing map data; and (3.) Update of the position of the base station in the map data.
  • the self-localization of the robot can be carried out in this case using an algorithm that determines the position of the robot on the map by comparing the data detected by the navigation sensor with the existing map data.
  • the former, known position of the base station can be used as a first localization hypothesis for the location of the robot. If, for example, in step 2 described above the localization of the robot fails, then the robot begins to compile a new map and the position of the base station is recorded in this new map. Upon completion of the robot's operational task, the user can be informed of the newly compiled map and can be asked whether it should be used to replace or augment the former map.
  • a robot may be employed in two or more mutually isolated areas of robot operation such as, for example, on various floors of a building.
  • a base station which the robot can definitively identify with the aid of its navigation sensor (as described above), may be located in each one of the areas of operation.
  • the robot “knows” which area of operation (e.g. on which floor) it is located in.
  • the navigation module of the robot can load the map data that is linked with the respective base station.
  • a prerequisite for this is that the user positions the robot in advance on one of the base stations or in the proximity of one of the base stations.
  • the robot can start a localization using the loaded map data.
  • the robot can use the position of the base station known from the map and its relative position to the same in order to accelerate the localization. This is done, for example, by using the position of the base station (or the docking position of the robot) as a localization hypothesis.
  • the area on the map in which the robot is attempting to localize itself is limited to an area surrounding the base station. This area is, for example, a square in front of the base station (and bordering on the same) or a circle around the base station. The lateral lengths/radius can depend on the distance between the robot and the base station.
  • the robot carries out the following steps for the purpose of localization: (1.) Exit the base station; (2.) Detect and identify the base station; (3.) Loading of map data linked with the base station; (4.) Localization of the robot on the map bases on the loaded map data; and (5.) Update of the position of the base station in the map data.
  • Step 1 is optional, depending on whether the robot is docked on the base station or not at the start.
  • a new map of the robot's area of operation is compiled. After the robot has completed its operational task, the user can be informed about the newly compiled map. The robot can ask the user whether the new map should be linked with the base station and permanently saved. In an alternative embodiment, the robot, in the event of a failed localization, can attempt a localization using the map data associated with other base stations. This may be useful, for example, if the user has replaced the base station with different one.
  • Navigation sensors such as, for example, sensors for distance measurement (see e.g. triangulation sensor in accordance with FIG. 2 ), are very sensitive measurement systems.
  • a disturbance such as, for example, an impact, can result in a decalibration of the sensor, leading to consistently erroneous measurements. This can significantly impair the navigation of the robot through its area of operation.
  • a basis station in accordance with the embodiments described here may be employed for this purpose (calibration), because it has well defined geometric features that are easily recognized by the navigation sensor.
  • a decalibration of the navigation sensor can be determined, for example, by measuring the distance, the length, the size and/or the angles of one or more geometric features (or more specifically, the thus derived navigation features) of the base station and by comparing these with a corresponding standard value. This standard value can also be used to recalibrate the navigation sensor.
  • individual geometric features of the base station may be designed such that their measurement allows for a direct inference to be drawn about at least one of the parameters of the navigation sensor that is to be calibrated.
  • numerous independent measurements may be combined, so that a measurement error of a single measurement will have less overall impact.
  • numerous independent geometric features of the base station such as, for example, width and depth, cf. FIG. 1B and FIG. 3 ) may be measured.
  • the same measurement may be carried out at different distances to the base station.
  • the travelled distance (odometry) between two measurement positions may be taken into consideration.
  • a navigation sensor in particular, a sensor for distance measurement
  • a triangulation sensor which emits structured light (see FIG. 2 , light beam 111 ) from a light source and takes a picture of the thus illuminated environment with a camera (see FIG. 2 , camera 112 ). Based on the camera picture and the relative position between the camera and the light source, the distance (see FIG. 2 , distance d) to an obstacle can be determined. Small changes in the relative position, caused by a shift (e.g. by approximately 1 ⁇ m) or a rotation (e.g.
  • FIG. 1A An example is shown in which, with the aid of a triangulation sensor (navigation sensor 110 ), the (horizontal) cross section of the base station 200 is detected.
  • FIG. 3C A possible result of this measurement can be seen in FIG. 3C . From this sensor data, various dimensions (dimension values) that allow for an inference to be drawn about the quality of the distance measurement can be derived.
  • the distance separating the points 1 and 4 can be determined and compared with the (known) actual width
  • the distance of one or more features on the front side of the base station (such as points 1, 2, 3 and 4) to the back wall can be determined and compared with the (known) actual depth of the base station;
  • the distance separating the robot 100 and the base station can also be determined, based, for example, on the width and/or depth of the base station 200 . This value may be used to calibrate a distance measurement.
  • the base station 200 can be its relatively small construction size (as compared with large furniture).
  • the navigation sensor should provide exact measurement results over the course of numerous meters, the base station itself is only a few centimeters wide and deep.
  • the light path inside of it can be extended. In this manner the accuracy of the calibration of a navigation sensor that emits targeted light can be improved.
  • the inside of the back wall can be mirrored. In such a case this would make the inner surface of the front housing wall visible to the navigation sensor 110 , meaning that the depth available as a measurement path will be twice that what it would be without the mirroring.
  • the following sensor parameters can be calibrated: Distance of the image sensor (e.g. CCD or CMOS sensor) from the lens (focal length); the distance of the optical axes from the light source (e.g. laser and lens); the inclination of the measurement plane (corresponds to the inclination of the optical axis of the light source, see FIG. 4A , Case b); zero point of the position x on the image sensor (see FIG. 2 ).
  • the last two mentioned parameters inclination of the optical axis of the light source and position of the zero point on the image sensor
  • greatly falsify the distance measurement especially at long distances, which results in systematic measurement errors.
  • Both of these triangulation sensor parameters can be calibrated, for example by measuring the width of the base station (or the distance between points 1 and 4 in FIG. 3C ) and by adapting (calibrating) the parameters such that measured value corresponds with a known reference value.
  • the depth of the base station e.g. the distance of point 3 from the back wall, see FIG. 3C ) is also known and may be used for the calibration of the mentioned parameters.
  • FIG. 4A shows a robot 100 with a navigation sensor 110 (in particular a triangulation sensor) that is carrying out a measurement of the cross section of the base station 200 .
  • the measurement is carried out in a plane that lies parallel to the floor at a distance to the floor h (Case a in FIG. 4A ), but it may also deviate from this (Case b in FIG. 4A ).
  • FIG. 4B shows a base station 200 that is comparable to the example of FIG. 3A , however with an opening 230 ′ (window) that is formed such that the position (and/or the extension) of a navigation feature is dependent on the distance to the floor (punctured line in FIG.
  • the point 2 as compared to Case as, is shifted to the left, which results in the distance between point 1 and point 2 in Case b being smaller, and the distance between the points 2 and 3 being greater, than in Case a.
  • the base station By measuring the base station at various distances it can be directly determined, for example, whether the sensor measurement is being carried out parallel to the floor surface or, if not, how large the tilt of the measurement plane is relative to a horizontal plane.
  • the third dimension of the base station is also used to directly determine and calibrate a sensor parameter (here the inclination of the optical axis of the light source or the measurement plane of the navigation sensor).
  • the base station requires a geometric feature (e.g. the position of point 2 or the distance between point 1 and 2) that definitively depends on the distance to the floor h.
  • the navigation module calculates a docking position from the detected geometric features (navigation features) of the base station and then directs the robot into this position. If the measurements of the navigation sensor are erroneous, the docking maneuver may fail, as then the calculated docking position will not correspond with the actually needed one.
  • the calculation of the docking position depends on one or more parameters that can be calibrated if the exact docking position is known. These parameters are, for example, the position of the navigation sensor 110 on the robot 100 , but also, for example, the shifting of a mirror of the optical reception device of the navigation sensor 110 .
  • the calibration can be carried out by means of a method based on trial and error.
  • the calculated docking position is varied and used to carry out the docking maneuver. This is done repeatedly and the rate of success is measured. The docking position with the highest rate of success is then used to calibrate the needed parameters. Varying the position (orienting) may occur randomly or systematically in small steps in an interval around the calculated position. Naturally, instead of the position the interesting parameters may also be directly varied and used to determine the interesting positions.
  • the robot 100 can determine its docking position and orientation based on the position of the base station 200 .
  • the navigation module can determine a path for the robot based on its calculated docking position and orientation and can direct the robot to this docking position. Nevertheless, the actual final position and final orientation of the robot 100 (at docking) may deviate from the planned and needed docking position and docking orientation due to odometry errors.
  • FIG. 5B shows that errors may occur merely due to small deviations, e.g. entailing that no contact between the charging contacts 220 of the base station and the corresponding charging contacts 120 of the robot will be established, thus impairing the autonomous functionality of the robot 100 .
  • the robot 100 In order to correct a small deviation in the position of the robot and/or its orientation from the actually needed docking position and/or orientation after completion of a docking maneuver, the robot 100 minimally varies its position (e.g. orientation). As shown in FIG. 5C , by means, for example, of a small rotation a charging contact can be established despite a less than exact docking position (see FIG. 5D ).
  • the robot for example, carries out a first rotation to the left by a predefined angle ⁇ . As depicted in FIG. 5C , this does not necessarily lead to success, which is why a second rotation in the opposite direction is carried out. In order to cover a symmetric area around the original orientation, the second rotation is carried out, for example, by an approximately twofold angle 2 ⁇ of that of the first rotation, however in the opposite direction.
  • Whether or not the needed docking orientation has been achieved can be determined, for example, by the presence of a voltage on the charging contacts or by means of a contact switch. If, for example, no successful contacting has been achieved, the robot can return to its beginning orientation after completing the described rotations.
  • the latter can be provided with one or more (flexibly mounted) nibs which engage in corresponding grooves on the robot.
  • these are formed, for example, by the two charging contacts 220 .
  • these nibs can slide into the corresponding grooves on the robot and thus define an exact docking position and orientation.
  • Connections for vacuuming out a dirt container on the robot by the base station or for refilling cleaning agents may fulfill a comparable function.
  • a certain docking area in front of the base station should be free of obstacles. This area should be large enough for the robot to be able to reliably recognize it when it passes by and it should afford enough space for a docking maneuver that is as simple and direct as possible. Such an area, for example, should afford a space on both the right and left side of the base station that is approximately equivalent to the diameter of the robot and a space in front of the base station approximately equivalent to twice that of the robot's diameter.
  • the influence of the user may result in various disturbances such as, for example, (i) the base station is so positioned that a wall runs through the docking area (in particular because the base station was placed too close in the corner of a room; (ii) small obstacles, such as the leg of a chair or scattered shoes are present in the docking area, blocking part of the approach path; (iii) small obstacles such as a cable or an article of clothing are left lying in the docking area that impede the movement of the robot, for example, by causing an increased slipping of the wheels.
  • the user will not intentionally cause these disturbances and will not intentionally place obstacles in the robot's path that impair access to the base station and thus a safe docking.
  • the attempt is made to detect such problems early and to then inform the user of them so that he may remove the disturbance.
  • the robot for example, is provided with a detection module that can recognize that calculating and/or travelling along the docking path has been significantly impaired or made impossible, for example, by one of the disturbances described above.
  • the navigation sensor for example, can be used for this by detecting obstacles in the docking area. Smaller obstacles in the docking area that the robot can drive over, for example, can be recognized by a sensor that detects an odometry error such as, for example, slipping.
  • the robot In order to inform the user of a recognized problem, the robot possesses at least one communication interface (also known as Human Machine Interface HMI). This may comprise a visual display directly on the robot or a sound signal, in particular a voice output.
  • HMI Human Machine Interface
  • This information includes, for example, the nature of the detected disturbance.
  • the robot can assess how severe the disturbance is and provide the user with this information (e.g. a minor problem (level I, interaction by the user not immediately needed), a relevant problem (level II, interaction by the user recommended/advisable), a grave problem (level III, interaction by the user required)).
  • level I a minor problem
  • level II a relevant problem
  • level III a grave problem
  • the user can decide how crucial it is to intervene.
  • the base station may have been placed too close to a wall, disturbing the docking procedure but not rendering it impossible (level I or II).
  • the user can decide that the base station should remain at its present position and the robot should attempt to dock.
  • the user may have accidentally placed the leg of a chair in front of the base station, rendering a direct docking maneuver impossible (level III).
  • This problem can be quickly remedied by the user, thus ensuring the functionality of the robot.
  • the user can report back to the robot via the communication interface. This report will state, for example, that; (i) the problem has been remedied; (ii) the problem is to be ignored and a docking attempted; (iii) the docking maneuver is to be postponed and the problem will be remedied later.
  • This classification of disturbances is, naturally, only of exemplary nature and may also be carried out in any other desired manner.
  • the robot In order that the user be informed of problems as soon as possible, the robot should always carry out the detection of disturbances immediately upon leaving the base station, hence in particular at the start of a new operational task.
  • the robot includes a robot controller that is generally programmable and is, depending on the application, programmed such that the robot can carry out the respective method.
  • the robot controller need not necessarily be realized in a single component of the robot. In general, all components that influence the externally visible behavior of the robot form a part of the robot controller.
  • the robot controller does not have to be physically installed within the mobile robot, but may also be partially located outside of the robot in stationary (control) devices such as, e.g. computers that are connected with the robot via a communication connection.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Aviation & Aerospace Engineering (AREA)
  • Remote Sensing (AREA)
  • General Physics & Mathematics (AREA)
  • Automation & Control Theory (AREA)
  • Electromagnetism (AREA)
  • Robotics (AREA)
  • Mechanical Engineering (AREA)
  • Control Of Position, Course, Altitude, Or Attitude Of Moving Bodies (AREA)
  • Manipulator (AREA)

Abstract

In the following, a system having an autonomous mobile robot and a base station for the robot is described. In accordance with one example, the robot comprises a navigation module with a navigation sensor for detecting geometric features of objects in the environment of the robot. The base station has at least one geometric feature which can be detected by the robot by means of the navigation sensor. The robot includes a robot controller that is coupled with the navigation module, the robot controller being configured to identify and/or localize the base station and/or to determine a docking position of the robot based on the at least one geometric feature of the base station.

Description

    CROSS REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation of U.S. patent application Ser. No. 15/757,245, dated Mar. 2, 2018, which is a § 371 National Phase of PCT/EP2016/070761, filed Sep. 2, 2016, which claims priority to German patent application DE 10 2015 114 883.8, filed Sep. 4, 2015, the disclosure which is incorporated herein by reference in their entireties.
  • TECHNICAL FIELD
  • The application relates to a method for the identification and localization of a base station of an autonomous mobile robot by the robot. It is further described how the base station may be employed for the calibration of a sensor of the robot and how the docking maneuver may be improved.
  • BACKGROUND
  • In recent years, autonomous mobile robots, in particular service robots, are being increasingly employed in the household, for example, for cleaning or to monitor a home. When these robots are not performing their tasks, they are generally docked to a base station. This allows, for example, the charging of a battery, the cleaning of a dirt collection container or the refilling of the robot's cleaning fluid to be carried out by the base station. In order for the robot to operate fully autonomously, it is important for it to be capable of reliably finding its way back to the base station. Various ways of achieving this described task are known. The publications US 2009 0281 661 A1 and US 2014 0100 693 A contain background information on this subject.
  • In general, once aspect consists in simplifying or improving known methods for identifying and locating a robot base station by the robot as well as the known methods for docking onto the base station.
  • SUMMARY
  • The problem described above is solved by the system , base station and method having the features and structures recited herein. Various embodiments and further developments of the present application are the subject matter of the independent claims.
  • In the following a system having an autonomous mobile robot and a base station for the robot will be described. In accordance with one example of the present disclosure, the robot comprises a navigation module with a navigation sensor for detecting the geometric features of objects in the environment of the robot. The base station has at least one geometric feature that can be detected by the robot by means of the navigation sensor. The robot includes a robot controller that is coupled to the navigation module and which is configured, based on the at least one geometric feature of the base station, to identify the base station and/or to localize and/or to determine a docking position of the robot.
  • A further example of the application relates to a base station for a mobile robot. The base station has a housing with at least one opening arranged in the housing which, due to its geometry, defines at least one geometric feature that is detectable by the sensor system of the robot.
  • Further, a method for an autonomous mobile robot will be described. In accordance with one example of the application the method comprises the detection of geometric features of objects in the environment of the robot by means of a navigation module of the robot which has a navigation sensor. At least one of the thus detected objects is a geometric feature of the base station. The method further comprises the identification and/or localization of the base station based on the at least one geometric feature of the base station.
  • In accordance with a further example of a system having an autonomous mobile robot and a base station, the robot comprises a navigation module with a navigation sensor for the detection of geometric features of objects in the environment of the robot. The base station has at least one geometric feature that can be detected by the robot by means of a navigation sensor. The navigation module is configured to test and/or calibrate the navigation sensor with the aid of the at least one detected geometric feature of the base station.
  • Further embodiments relate to methods for an autonomous mobile robot. In accordance with one example a method comprises the detection of geometric features in the environment of robot by means of a navigation module of the robot which has a navigation sensor, wherein at least one of the detected features is a geometric feature of the base station. The navigation sensor is calibrated and/or tested with the aid of the at least one geometric feature of the base station.
  • A further method serves to dock an autonomous mobile robot onto a base station. In accordance with one embodiment, the method comprises the detection of a docking position of the robot on the base station, wherein the docking position comprises a location and an orientation of the robot, as well as the navigation of the robot into the docking position. Afterwards it is tested whether the robot has correctly docked onto the base station. If this is not the case, the position of the robot is varied and it is once again tested whether the robot has correctly docked onto the base station. The varying and testing is carried out until either the test is successful or an abort criterion is fulfilled.
  • In accordance with a further embodiment a method for the automatic docking of an autonomous mobile robot onto a base station includes the detection of obstacles by means of a navigation module of the robot which has a navigation sensor, as well as testing whether, within a defined area around the base station, access to the base station for the robot is impeded by detected obstacles. If the test reveals that access to the base station is impeded for the robot, a disturbance is communicated via a user interface.
  • A further exemplary method for an autonomous mobile robot includes the detection of geometric features of objects in the environment of the robot by means of a navigation module of the robot which comprises a navigation sensor, as well as the navigation of the robot based on at least one of the detected geometric features and on an electronic map of the area in which the robot operates. The position of a base station of the robot is designated on the electronic map. The method further includes testing whether the detected geometric features contain a geometric feature that is linked with the base station. If this is the case, the current position of the base station is determined based on the geometric feature that is linked with the base station and the position of the base station is updated on the electronic map. Alternatively, the base station may be detected and located in a different manner in order to update its position on the map of the robot.
  • A further example of a method for an autonomous mobile robot includes the detection of geometric features of objects in the environment of the robot by means of a navigation module of the robot which comprises a navigation sensor, as well as the navigation of the robot based on at least one of the detected geometric features and on an electronic map of the area in which the robot operates. The position of a base station of the robot is designated on the electronic map. In accordance with the method, a first geometric feature that is not defined by the base station is linked with the position of the base station. This first geometric feature is tracked using a SLAM algorithm, wherein a position of the first geometric feature on the electronic map is kept current and the position of the base station is saved as a relative position relative to the position of the first geometric feature.
  • BRIEF DESCRIPTION OF THE FIGURES
  • In the following, the present disclosure is described in greater detail by means of the examples shown in the figures. The figures are not necessarily to scale and the application is not limited to only the shown aspects. Instead emphasis is placed on illustrating the underlying principles of the application. The figures show:
  • FIG. 1 shows a robot with a base station in its area of robot operation.
  • FIG. 2 schematically shows by means of a drawing the optical distance measurement using triangulation.
  • FIG. 3 shows an example of a base station with geometric features that are detectable by the navigation sensor of the robot and which are defined by openings in the front housing wall of the base station.
  • FIG. 4 illustrates the detection of the geometric figures of a base station using the navigation sensor of the robot, as well as possible systematic measurement errors.
  • FIG. 5 illustrates a method for docking a robot onto a base station, wherein the robot continues to vary its orientation until it has correctly docked.
  • DETAILED DESCRIPTION
  • In general, a mobile robot should be able to dock onto its base station reliably and safely. Known systems (base stations and robots) and methods for finding and recognizing a base station and its exact position and orientation often employ special sensors in the robot (e.g. a light beam), complex image processing algorithms, which must be executed in addition to the navigation algorithms, and/or special markings on the base station or in the area in which the robot is operated. Further, the locating of the base station and the docking of the robot onto the base station should be robust against many different kinds of disturbances. For example, shifting of the base station may impair a reliable localization of the base station. The functionality of the sensors employed by the robot for navigation cannot be reliably tested in an unknown environment, such as in the area in which the robot is operated. The approach to the base station (docking maneuver) may sometimes end in a failed charging contact due to odometry errors. The user, for example, may place obstacles near the base station that disrupt the docking maneuver and thus result in a failed docking.
  • In view of the deficits of common robot base station systems and methods for identifying and localizing a base station, as well as for the reliable docking of the robot onto a base station that were described above, a need exists for improvements. For example, it would be desirable to make the localization of the base station of the robot possible using a navigation sensor already present in or on the robot, without the need for the base station to send out a signal or for special markings to be attached to it. In accordance with some of the embodiments described here, for this purpose the base station is recognized and localized with the aid of its geometric shape (e.g. certain geometric characteristics (outer form/shape) of the housing of the base station). This means that the base station itself exhibits, due to its outer form/shape, at least one geometric feature, with the aid of which the robot may recognize the base station. Special markings on the housing of the base station (which nevertheless have no influence on its geometric form/shape) or the emission of a (locating) signal are no longer necessary. In general, a robot should quickly and reliably find its way back to its base station after completion of its (e.g. cleaning) task. In accordance with some of the embodiments described here, for this purpose the position of the base station is marked on a map and is kept current by means of various methods. A further desirable capability of the robot is that it be able to maintain its bearings (to accurately and quickly orient itself) in different areas in which the robot operates (such as, for example, on the various floors of a building). In accordance with some of the embodiments described here, for this purpose one base station is employed in each of the areas in which the robot operates that the robot is able to distinguish from one another. The robot links a map of the respective area of operation with each base station. Once the robot has identified a specific base station, it can immediately begin a (self) localization with the aid of the map that is linked with each base station.
  • A further desirable capability of a mobile robot is a robust and exact docking onto the base station so that, for example, a contact between the charging contacts of the robot and those of the base station can be reliably carried out. In accordance with some of the embodiments described here, for this purpose, after reaching a previously calculated docking position, the position and orientation of the robot is slightly corrected in order to be able to safely dock. A docking should also be possible in the case of possible disturbance (e.g. access to the base station is blocked). In accordance with some of the embodiments described here, for this purpose a possible disturbance is detected and the user is informed of it.
  • In general it would be desirable for it to be possible to test the functionality of a navigation sensor of the robot within its area of operation and, if necessary, to recalibrate it. In accordance with one of the embodiments described here, for this purpose a base station with its a-priori known geometric parameters is employed.
  • FIG. 1A shows an autonomous mobile robot 100 and a corresponding base station 200 in an operation area of the robot. The robot has a navigation module with at least one navigation sensor 110 for orienting itself in its environment and for navigating throughout the operation area of the robot, thus enabling it to autonomously carry out a task. The navigation module of the robot 100 collects with the aid of the navigation sensor 110 information regarding the position of “navigation features” in the environment of the robot, e.g. geometric features of objects such as, e.g. those of obstacles, as well as information regarding the floor covering, etc. Geometric features are, for example, surfaces (e.g. a wall), lines (e.g. the contours of a wall, furniture or other objects in the environment of the robot 100) and points (e.g. on corners or edges of objects in the environment of the robot 100). Every navigation feature can be associated with a position (including orientation) in the room and, if needed, recorded on the electronic map of the robot 100. The navigation module operates, for example, using an obstacle avoidance strategy and/or a SLAM algorithm (Simultaneous Localization and Mapping) and/or one or more maps of the area in which the robot operates. The robot can create a new map of the robot operation area while in operation or it can make use of an already existing map before beginning its task. An already existing map can be created by the robot itself during the completion of an early task, such as an exploratory run, or may be provided by another robot and/or a person.
  • The navigation sensor 110 detects information about the environment of the robot (and thus about the area in which it operates), in particular about its geometric features in one, two or three dimensions. The navigation sensor 110 may be, for example, a sensor for distance measurement such as, for example, an optical and/or acoustic sensor that detects the distance between the sensor and an obstacle by means of triangulation or by measuring the travel time of an emitted signal (e.g. a laser beam or an ultrasonic signal). Accordingly, e.g. triangulation sensors, time of flight cameras, laser scanners, ultrasonic sensors and the like may be employed as navigation sensors. With the aid of this navigation sensor 110, the navigation module of the robot 100 can determine the distance between the robot and the individual points, lines and/or surfaces of objects in the environment of the robot. The thus detected points, lines and/or surfaces are designated as “navigation features” and are saved on a map of the robot (i.e. the geometry of the navigation feature and its position, including orientation, is designated on the electronic map). Later, the robot can orient itself with the aid of these navigation sensors and/or avoid the detected obstacles, thus preventing collision. Another typical example of a navigation sensor 110 is a camera (e.g. with a CCD or CMOS image sensor) that has a detection module for identifying the corners and edges of object (navigation features) in the environment recorded on an image by means of image data processing. In this manner the position of a navigation feature within a projected plane relative to the robot can be determined. When the robot moves, the position, including the orientation, of the navigation feature and that of the robot within the room can be determined on the basis of this movement. This procedure is known as visual SLAM.
  • FIG. 1C shows a simple example of a base station 200. It consists of a main body 210 with geometric features that can be detected by the navigation sensor as navigation features. The main body 210 of the base station 200 may include various further components such as, for example, components for charging the battery of the robot 100 or for vacuuming out dirt. Two charging contacts 220, for example, are arranged on the base station 200. Corresponding contact are arranged on the robot 100 which must come into contact with the charging contacts 220 in order to allow a successful charging of the robot battery. This means that, in order to charge its batter, the robot must dock onto the base station at a certain position and in a certain orientation. The position and orientation required for this (and, thus, the path to be followed), can be determined by the robot when the position and orientation of the main body 200 are known to it.
  • In accordance with the embodiments described here, the geometric characteristics of the base station 200 (i.e. its outer form/shape or its individual parts) are detected as navigation features with the aid of the navigation sensor 110 of the robot for the purpose of its navigation. By determining and interpreting the characteristics of the navigation features (with the aid of the navigation module of the robot), the navigation features of the base station, and thus the base station itself, can be definitively and without a great deal of additional effort identified, their position and orientation in the robot's environment can be determined and they can be recorded on the map of the robot. Simple criteria are used to achieve this purpose such as, for example, (1.) the distance separating certain points (e.g. corners) of a feature from other features, (2.) the length of edges/segments/lines, (3.) the size of surfaces, (4.) the location of features relative to each other, in particular the angle between two edges, segments or lines or the angle defined by three specific points (corners), (5.) ratios (relative sizes, lengths), e.g. the ratio between the lengths of two lines, and (6.) error metrics, such as the (squared) deviation of the (erroneous) measurement from the construction-related standard values (with the aid of a larger quantity of points, a straight line, for example, can be determined that can be established by means of linear regression; the degree of deviation of a specific point from the regression lines can serve as a criterion as to whether the given point altogether belongs to the assumed line).
  • For example, at least some of the detected features designate a width, a depth and/or a height of a part of the main body of the base station. The criteria used to locate the base station should be easy to calculate and be robustly determinable from various positions. This is explained in greater detail in the following.
  • In the example shown in FIG. 1A, the distance to obstacles in the environment of the robot is determined, e.g. by means of optic triangulation, for the purpose of which structured light (in a visible or non-visible range) in the form of a line running approximately parallel to the surface of the floor is emitted. The principle of optical triangulation is illustrated in FIG. 2. The emitted structured light 111 strikes an obstacle H and is diffusely scattered off of it. A camera takes a picture of the environment that depicts the light that is reflected back from the obstacle (e.g. a light line). By means of triangulation, the distance d to the illuminated obstacles H at selected points or along the entire line can be determined based (in the case of a horizontal light line) on the vertical position x of the structured, reflected light (i.e. of the light line) in the picture.
  • In order to ensure good detection, the base station 200 is geometrically designed such that, because of its height, it is illuminated by the light emitted from the navigation sensor 110 and, because of its depth, is clearly distinguishable from a wall. Furthermore, the surface of the base station is designed such that the light emitted from the navigation sensor 110 is reflected in a well diffused state (meaning that it exhibits no absorbing or mirroring surfaces).
  • FIG. 1B shows, as an example, the possible results of a sensor measurement of the scenario shown in FIG. 1A. In particular, the robot 100 recognizes with the aid of its navigation modules and its navigation sensors 110 two line shaped navigation FIGS. 201 and 202 stemming from the measurement of the base station. In the following, the navigation features 201, 202 will be designated as boundary lines or, in short, as lines. These two lines 201 and 202 each have a characteristic length which corresponds to the width (in horizontal direction) of the side surface (line 201) and of the front surface (line 202). Further, the two lines 201 and 202 include a certain angle (e.g. a right angle). If, for example, the base station 200 (e.g. in a horizontal plane that lies at a certain height above the floor) has a rectangular cross sectional area having side lengths of, e.g. 5 cm and 15 cm and the robot 100, with the aid of its navigation sensor 110, detects two almost rectangular lines having lengths of nearly 5 cm (line 201) and 15 cm (line 202), then the robot 100 can interpret these lines as the base station. In addition to this, use may be made of the side ratio of 5:15=1:3. As illustrated in FIG. 1B, one side of the base station cannot be recognized by the robot due to the shadowing caused by the front surface of the navigation sensor. In order to detect this additional geometric feature of the base station as well, the robot can move round the base station until it has also reliably recognized the second side surface. Thus, in this case, the base station will be regarded from at least two different positions.
  • In the methods described above, the base station 200 is recognized solely based on its rectangular form. In such cases, the probability for an erroneous detection is high, as any rectangular object (for example, a box) having the corresponding side lengths would be recognized as the base station. Furthermore, the described corners will have generally been replaced by rounded edges, both for safety reasons (risk of injury), as well as for reasons of design. In principle, such rounded edges are also detectable with the aid of the navigation sensor described above. The calculation effort needed for the processing of rounded forms, however, is greater than that needed for straight lines.
  • For this reason, in accordance with the embodiment described here, the inside of the base station is also utilized for its detection. For this purpose, one or more openings are introduced into the main body of the base station. Individual components (for example, for vacuuming dirt) of the base station may be visible through these openings. This may also be omitted, however, so that only one or more inner walls can be recognized.
  • FIG. 3A shows an example of a base station 200 with rounded edges and two openings 230 arranged in the front side of the base station. FIG. 3B shows a cross section view through the base station 200 of FIG. 3A, wherein the section plane is a horizontal plane that lies at a certain height above the floor. This certain height is the height at which the navigation sensor 110 carries out the distance measurement to obstacles in the environment of the robot. FIG. 3C shows an example of the results of such a measurement when the robot is standing at a certain distance (for example 0.5-1 m) in front of the base station. From here it can see parts of the back wall of the base station 200, which all lie along one line. In addition to this, the robot can determine the position of four points (1, 2, 3, 4) with negligible measurement errors. The position and size of the two openings 230 (windows) has been chosen in this case such that the distance separating any of the two points is different for each pair of points, (meaning, in particular, that d(1; 2)≠d(1; 3)≠d(1; 4)≠d(2; 3)≠d(2; 4)≠d(3; 4), wherein d(i, j) designates the distance between point i and point j). These distances can be easily calculated and compared to stored standard values. The distance separating the points from the line formed by the back wall may also be compared, for example. The ratios of the distances to each other may also be taken into consideration. By using an error function such as, for example, the coefficient of determination of a regression model, the degree of exactness with which the points 1, 2, 3, and 4 lie along a line parallel to the back wall can be tested. By these means numerous features are made available that significantly reduce the probability of an erroneous detection, so that in daily practice it will virtually not occur.
  • To protect the inside from contamination, a cover 250, made of a transparent material that allows the light emitted from the navigation sensor 110 to pass through, may be arranged over the openings 230. The emitted light may lie, for example, in an (infrared) spectrum that is not visible to humans, so that the cover as viewed by the robot is transparent, but, as viewed by humans, will appear colored and nontransparent. Such a cover 250 should not only cover the openings 230, but should also, independent of this, be used to form differing geometric shapes that can be recognized by the user and the robot. In this manner, aspects of design may be combined with a simplified detectability. The cover 250 may be provided with an antireflective coating that is matched to the wavelength of the light emitted by the navigation sensor 110.
  • In the case of the described measurement in a (horizontal) plane, the latter may slightly vary in its distance to the floor (e.g. due to differing installation heights of the navigation sensor 110 or when the plane of measurement is not completely parallel to the floor, e.g. because the robot is standing somewhat askew). In order that the base station may be reliably recognized in this plane based on its geometric form, at least a selection of the recognizable navigation features (or more specifically, their underlying dimensions) must be found in the environment of the expected cross section measurement, irrespective of the actual height of measurement (the distance of the plane in which the distance measurement is carried out from the floor).
  • In some applications, the robot will be employed in two mutually isolated areas of robot operation such as, for example, on two different floors of a building. In each of the areas of operation a base station 200 is located that the robot 100 can definitively recognize with the aid of its navigation sensor 110. If the robot 100 is thus now capable of differentiating between the base stations 200, it also immediately obtains information as to which of the various areas of operation it is located in without the need for the user to additionally intervene.
  • In order that the base stations 200 be distinguishable, some parts of the geometric form of the base station, for example, may be altered by the user (e.g. at the time the base station is put into operation), meaning that a feature that is recognizable by the navigation sensor 110 may be altered in a previously defined manner. This alteration can be achieved, for example, by shifting, removing or adding parts to the base station. In the case of the exemplary base station 200 in FIG. 3A, an alteration can be carried out by changing the size (width) of one of the two openings, for example. Thus, for example, the location of point 3 (FIG. 3C) can be varied by adding, for example, an additional screen. For example, a sliding screen may be arranged behind the centre panel between the two viewing windows.
  • Autonomous mobile robots 100 that mark their base stations 200 on a map are basically known. Often the base station serves as the starting point of the map, i.e. the base station has a permanent position on the map. The position of the base station, however, can change. Detecting the base station based on characteristics (navigation features) that can be detected by the navigation sensor 110 makes it possible to easily update the position of the base station on the map. However, it is also possible to detect the base station by other means, thereby allowing the position of the base station on the map of the robot to be updated. The change in the position of the base station may be carried out by a user, e.g. (1.) while the robot is in operation or (2.) while the robot is docked on the base station. Furthermore, due to measurement and odometry errors, the mapping carried out by the robot and/or its localization on the map (SLAM) may lead to the result that the position of the base station anticipated by the robot does not correspond to its actual position. In the worst case, this can result in the loss of all (meaningful) information about the position of the base station. In the following, examples will be used to explain how this problem can be solved, or at least mitigated.
  • Position updates using a SLAM algorithm—SLAM algorithms (Simultaneous Localization and Mapping), by means of which the position of the robot and selected navigation features that have been detected by the navigation sensor are continuously correlated, are often employed for the navigation of autonomous mobile robots. This enables to robot to carry out a robust mapping despite measurement and odometry errors. SLAM demands a relatively high amount of computing capacity, which is why only a few selected and readily visible navigation features such as, for example, walls are taken into consideration in order to limit the needed calculation effort. The positions of the navigation features (and those of the robot) that are tracked in the SLAM algorithm are continuously corrected; meaning they are at no permanent (relative) position in relation to other objects that are not tracked in the SLAM algorithm. The features of the base station are generally too small to be tracked by the SLAM algorithm and they may therefore shift their recorded position relative to a wall (which is tracked by the SLAM algorithm). This can result in the recorded position being located in the wrong room or even outside of the area of operation. For this reason it is useful to update the position of the base station.
  • Since the base station also has detectable geometric features (navigation features), the position of the base station can be easily kept current using the SLAM algorithm. For this purpose, for example, at least one easily detectable feature of the base station (cf. FIG. 3) is selected and tracked with the aid of the SLAM algorithm. This can be, for example, the segment formed by points 2 and 3 in FIG. 3C. As an alternative, a feature can also be used that has been detected in the proximity of the base station. This can be, for example, the line 300 in FIG. 1B, which is formed by the wall against which the base station stands. In both cases the robot records which of the features tracked by the SLAM algorithm determines the position of the base station. In addition to this, the robot can save the position of the base station relative to this navigation feature. When the robot wants to return to the base it can move to a position near this feature, thus ensuring that it will reliably find its way back again.
  • Position updates during operation of the robot—In the event that the base station, for example, is moved by the user while the robot is in operation, it is advantageous for the robot 100 to recognize the base station 200 when moving past it and to update the map with the new position. Since the navigation sensor detects features of the base station to navigate the robot, it is sufficient in this case for the newly detected features to be tested as to whether they might be part of the base station. In this case, as a first step, an easily recognized feature, for the detection of which only a small amount of additional calculation time is required, can be used. After recognition has been confirmed, further characteristics of the base station can be tested.
  • For example, in the case of the base station in FIG. 3A, segments (e.g. lines of a certain length and orientation) can be looked for that correspond in their length to the distance separating points 2 and 3 in the FIG. 3C. After such a segment has been found, it can be further tested as to whether a back wall is recognizable, whether this is at the correct distance, and whether the points 1 and 4 are present and lie along one line. While doing so, the length of recognized segments, for example, can be routinely detected by the navigation module in order to test them regarding their relevance for the navigation.
  • Once the base station has been recognized at a new position, various options for further action become available to the robot. If the new position deviates only slightly from the former one (for example, by a distance of less than 1 m), then the new position of the base station is recorded and the former one is deleted. This is particularly useful when the navigation sensor simultaneously examines the former position of the base station and no base station is detected at this position. If a base station is detected at a location in the area of robot operation that is far away from the previous base station, then this may be due to the presence of a second base station. In this case the position of the base station is newly recorded and the former position is retained for later verification. When the robot is in the proximity of the former position, it tests whether the base station is still present. Depending on the result, the former position is either deleted or it is recorded that there are two base stations located in the area of operation of the robot. The respective positions (including orientation) of both base stations can be saved on the map.
  • It can also occur that the robot, according to the information on its map, finds itself in the proximity of a base station but does not detect it. This may be the case if the position of the base station has been changed by the user, but also if the user has only temporarily removed it, for example to clean it. In the simplest case the robot will delete the position of the base station from the map data. Alternatively, this position on the robot's map may be marked as “questionable”. If while carrying out its operational task the robot later recognizes a base station at a different position, then the position marked as questionable is deleted and the new position of the base station is recorded. In the opposing case, the robot returns to the position marked as questionable when it wants to return to the base station. If again no base station is found at this position, then a special search program is started to find the base station, wherein the robot moves around the area of operation and looks specifically for features (navigation features) of the base station. In addition to this, a message can be sent to the user (e.g. by means of a user interface, e.g. an application on a tablet computer or mobile telephone), allowing the user to react to the situation, if needed.
  • The measurement accuracy of the navigation sensor 110 may deteriorate as the distance to the object to which the distance measurement is to be carried out increases. In the case of object with elongated dimensions, the accuracy also depends on their orientation. This may particularly be the case when triangulation sensors are used. If, for example, the base station 200 of FIG. 3 is viewed directly from the front, then the distance separating the points 2 and 3 can be determined at a distance of approximately 2 m with sufficient accuracy. If the base station, however, is viewed from an angle of 45°, then, in order to ensure a reliable measurement, the distance between the base station 200 and the robot 100 should not exceed about 1 m. Based on this, a maximum distance for testing whether a detected navigation feature belongs to the base station can be determined in order to improve the quality of the detection of the base station. The maximum distance for carrying out a test may also depend on the orientation of the feature (relative to the robot).
  • Position updates at the start of robot operation—A user may move the robot together with the base station while the robot is docked onto it and while the robot, for example, is in pause modus. If the robot is using an existing map that, for example, it had compiled while carrying out a previous operational task, then its new position (and that of the base station) will no longer correspond with the existing navigation information. In accordance with the embodiment described here, the robot may carry out the following steps at the start of its operational task: (1.) Exit the base station and collect map data that is relevant for its operational task; (2.) Localization of the robot in the existing map data; and (3.) Update of the position of the base station in the map data.
  • The self-localization of the robot can be carried out in this case using an algorithm that determines the position of the robot on the map by comparing the data detected by the navigation sensor with the existing map data. In order to accelerate the localization algorithm, the former, known position of the base station can be used as a first localization hypothesis for the location of the robot. If, for example, in step 2 described above the localization of the robot fails, then the robot begins to compile a new map and the position of the base station is recorded in this new map. Upon completion of the robot's operational task, the user can be informed of the newly compiled map and can be asked whether it should be used to replace or augment the former map.
  • Navigation in numerous areas of robot operation—A robot may be employed in two or more mutually isolated areas of robot operation such as, for example, on various floors of a building. A base station, which the robot can definitively identify with the aid of its navigation sensor (as described above), may be located in each one of the areas of operation. As a result of this identification of the base station, the robot “knows” which area of operation (e.g. on which floor) it is located in. Thus the navigation module of the robot can load the map data that is linked with the respective base station. A prerequisite for this is that the user positions the robot in advance on one of the base stations or in the proximity of one of the base stations. The robot can start a localization using the loaded map data. When doing so, as described above, the robot can use the position of the base station known from the map and its relative position to the same in order to accelerate the localization. This is done, for example, by using the position of the base station (or the docking position of the robot) as a localization hypothesis. In an alternative form, the area on the map in which the robot is attempting to localize itself is limited to an area surrounding the base station. This area is, for example, a square in front of the base station (and bordering on the same) or a circle around the base station. The lateral lengths/radius can depend on the distance between the robot and the base station.
  • In accordance with a further embodiment, the robot carries out the following steps for the purpose of localization: (1.) Exit the base station; (2.) Detect and identify the base station; (3.) Loading of map data linked with the base station; (4.) Localization of the robot on the map bases on the loaded map data; and (5.) Update of the position of the base station in the map data. Step 1 is optional, depending on whether the robot is docked on the base station or not at the start.
  • If no map data is linked with the recognized base station or the localization on the respective map fails, a new map of the robot's area of operation is compiled. After the robot has completed its operational task, the user can be informed about the newly compiled map. The robot can ask the user whether the new map should be linked with the base station and permanently saved. In an alternative embodiment, the robot, in the event of a failed localization, can attempt a localization using the map data associated with other base stations. This may be useful, for example, if the user has replaced the base station with different one.
  • Navigation sensors such as, for example, sensors for distance measurement (see e.g. triangulation sensor in accordance with FIG. 2), are very sensitive measurement systems. A disturbance such as, for example, an impact, can result in a decalibration of the sensor, leading to consistently erroneous measurements. This can significantly impair the navigation of the robot through its area of operation. In order to ensure the reliable functioning of the robot it may be necessary to systematically recognize measurement errors of the navigation sensors and, if possible, compensate them by means of calibration. A basis station in accordance with the embodiments described here may be employed for this purpose (calibration), because it has well defined geometric features that are easily recognized by the navigation sensor.
  • A decalibration of the navigation sensor can be determined, for example, by measuring the distance, the length, the size and/or the angles of one or more geometric features (or more specifically, the thus derived navigation features) of the base station and by comparing these with a corresponding standard value. This standard value can also be used to recalibrate the navigation sensor. In addition to this, individual geometric features of the base station may be designed such that their measurement allows for a direct inference to be drawn about at least one of the parameters of the navigation sensor that is to be calibrated.
  • In order to improve the accuracy of the calibration, numerous independent measurements may be combined, so that a measurement error of a single measurement will have less overall impact. For this purpose, numerous independent geometric features of the base station (such as, for example, width and depth, cf. FIG. 1B and FIG. 3) may be measured. Alternatively or additionally, the same measurement may be carried out at different distances to the base station. In particular, the travelled distance (odometry) between two measurement positions may be taken into consideration.
  • One example of a navigation sensor (in particular, a sensor for distance measurement) is a triangulation sensor, which emits structured light (see FIG. 2, light beam 111) from a light source and takes a picture of the thus illuminated environment with a camera (see FIG. 2, camera 112). Based on the camera picture and the relative position between the camera and the light source, the distance (see FIG. 2, distance d) to an obstacle can be determined. Small changes in the relative position, caused by a shift (e.g. by approximately 1 μm) or a rotation (e.g. by approximately 0.01°) of the light source (and thus of the light beam 111) relative to the camera 112 (brought about, for example, by impact) may result in systematic errors in the distance measurement that lead to a decalibration of the (previously calibrated) navigation sensor 110. An example is shown in FIG. 1A in which, with the aid of a triangulation sensor (navigation sensor 110), the (horizontal) cross section of the base station 200 is detected. A possible result of this measurement can be seen in FIG. 3C. From this sensor data, various dimensions (dimension values) that allow for an inference to be drawn about the quality of the distance measurement can be derived. For example, (a) the distance separating the points 1 and 4 can be determined and compared with the (known) actual width, (b) the distance of one or more features on the front side of the base station (such as points 1, 2, 3 and 4) to the back wall can be determined and compared with the (known) actual depth of the base station; (c) it can be tested whether the points 1, 2, 3 and 4 lie along one line, and/or (d) the angle between the back wall and the line defined by the front features ( points 1, 2, 3 and 4) can be determined (in the present example, this angle should ideally be zero, i.e. the line is parallel to the back wall). In particular, the distance separating the robot 100 and the base station can also be determined, based, for example, on the width and/or depth of the base station 200. This value may be used to calibrate a distance measurement.
  • One particular challenge for the accuracy of a calibration using the base station 200 can be its relatively small construction size (as compared with large furniture). Although the navigation sensor should provide exact measurement results over the course of numerous meters, the base station itself is only a few centimeters wide and deep. By arranging mirrors within the base station the light path inside of it can be extended. In this manner the accuracy of the calibration of a navigation sensor that emits targeted light can be improved. For example, in the base station in accordance with FIG. 3A, the inside of the back wall can be mirrored. In such a case this would make the inner surface of the front housing wall visible to the navigation sensor 110, meaning that the depth available as a measurement path will be twice that what it would be without the mirroring.
  • In the case of a triangulation sensor, the following sensor parameters can be calibrated: Distance of the image sensor (e.g. CCD or CMOS sensor) from the lens (focal length); the distance of the optical axes from the light source (e.g. laser and lens); the inclination of the measurement plane (corresponds to the inclination of the optical axis of the light source, see FIG. 4A, Case b); zero point of the position x on the image sensor (see FIG. 2). In particular the last two mentioned parameters (inclination of the optical axis of the light source and position of the zero point on the image sensor) can, in the event of a decalibration, greatly falsify the distance measurement, especially at long distances, which results in systematic measurement errors. Both of these triangulation sensor parameters can be calibrated, for example by measuring the width of the base station (or the distance between points 1 and 4 in FIG. 3C) and by adapting (calibrating) the parameters such that measured value corresponds with a known reference value. The depth of the base station (e.g. the distance of point 3 from the back wall, see FIG. 3C) is also known and may be used for the calibration of the mentioned parameters.
  • Whereas, in accordance with the previously described example, only the form of the base station in a horizontal sectional plane is taken into consideration, alternatively or additionally to this navigation features that are dependent on the distance to the floor of the (horizontal) measurement plane may also be considered.
  • FIG. 4A shows a robot 100 with a navigation sensor 110 (in particular a triangulation sensor) that is carrying out a measurement of the cross section of the base station 200. In the ideal case, the measurement is carried out in a plane that lies parallel to the floor at a distance to the floor h (Case a in FIG. 4A), but it may also deviate from this (Case b in FIG. 4A). FIG. 4B shows a base station 200 that is comparable to the example of FIG. 3A, however with an opening 230′ (window) that is formed such that the position (and/or the extension) of a navigation feature is dependent on the distance to the floor (punctured line in FIG. 4B) at which the sensor measurement is carried out. FIG. 4C represents, as navigation features, the corner points of the openings 230 and 230′ of the base station for Case a (distance to floor of the measurement h=ha), and FIG. 4D represents, as navigation features, the corner points of the openings 230 and 230′ of the base station for Case b (distance to floor of the measurement h=hb). In Case b, the point 2, as compared to Case as, is shifted to the left, which results in the distance between point 1 and point 2 in Case b being smaller, and the distance between the points 2 and 3 being greater, than in Case a. By measuring the base station at various distances it can be directly determined, for example, whether the sensor measurement is being carried out parallel to the floor surface or, if not, how large the tilt of the measurement plane is relative to a horizontal plane. Thus, with this procedure, in addition to the two-dimensional cross section measurement, the third dimension of the base station is also used to directly determine and calibrate a sensor parameter (here the inclination of the optical axis of the light source or the measurement plane of the navigation sensor). For this, the base station requires a geometric feature (e.g. the position of point 2 or the distance between point 1 and 2) that definitively depends on the distance to the floor h.
  • In accordance with a further embodiment, the navigation module calculates a docking position from the detected geometric features (navigation features) of the base station and then directs the robot into this position. If the measurements of the navigation sensor are erroneous, the docking maneuver may fail, as then the calculated docking position will not correspond with the actually needed one. The calculation of the docking position depends on one or more parameters that can be calibrated if the exact docking position is known. These parameters are, for example, the position of the navigation sensor 110 on the robot 100, but also, for example, the shifting of a mirror of the optical reception device of the navigation sensor 110. The calibration can be carried out by means of a method based on trial and error. For this purpose, the calculated docking position is varied and used to carry out the docking maneuver. This is done repeatedly and the rate of success is measured. The docking position with the highest rate of success is then used to calibrate the needed parameters. Varying the position (orienting) may occur randomly or systematically in small steps in an interval around the calculated position. Naturally, instead of the position the interesting parameters may also be directly varied and used to determine the interesting positions.
  • As described further above, the robot 100 can determine its docking position and orientation based on the position of the base station 200. In accordance with the example illustrated in FIG. 5A, the navigation module can determine a path for the robot based on its calculated docking position and orientation and can direct the robot to this docking position. Nevertheless, the actual final position and final orientation of the robot 100 (at docking) may deviate from the planned and needed docking position and docking orientation due to odometry errors. FIG. 5B shows that errors may occur merely due to small deviations, e.g. entailing that no contact between the charging contacts 220 of the base station and the corresponding charging contacts 120 of the robot will be established, thus impairing the autonomous functionality of the robot 100.
  • In order to correct a small deviation in the position of the robot and/or its orientation from the actually needed docking position and/or orientation after completion of a docking maneuver, the robot 100 minimally varies its position (e.g. orientation). As shown in FIG. 5C, by means, for example, of a small rotation a charging contact can be established despite a less than exact docking position (see FIG. 5D).
  • Beginning in the first docking position in accordance with FIG. 5B, the robot, for example, carries out a first rotation to the left by a predefined angle α. As depicted in FIG. 5C, this does not necessarily lead to success, which is why a second rotation in the opposite direction is carried out. In order to cover a symmetric area around the original orientation, the second rotation is carried out, for example, by an approximately twofold angle 2α of that of the first rotation, however in the opposite direction.
  • Whether or not the needed docking orientation has been achieved can be determined, for example, by the presence of a voltage on the charging contacts or by means of a contact switch. If, for example, no successful contacting has been achieved, the robot can return to its beginning orientation after completing the described rotations.
  • In order to further increase the robustness of the docking maneuver onto the base station, the latter can be provided with one or more (flexibly mounted) nibs which engage in corresponding grooves on the robot. In FIG. 1C these are formed, for example, by the two charging contacts 220. By means of the rotation of the robot, these nibs can slide into the corresponding grooves on the robot and thus define an exact docking position and orientation. Connections for vacuuming out a dirt container on the robot by the base station or for refilling cleaning agents may fulfill a comparable function.
  • In order to ensure a successful docking maneuver, a certain docking area in front of the base station should be free of obstacles. This area should be large enough for the robot to be able to reliably recognize it when it passes by and it should afford enough space for a docking maneuver that is as simple and direct as possible. Such an area, for example, should afford a space on both the right and left side of the base station that is approximately equivalent to the diameter of the robot and a space in front of the base station approximately equivalent to twice that of the robot's diameter.
  • The influence of the user may result in various disturbances such as, for example, (i) the base station is so positioned that a wall runs through the docking area (in particular because the base station was placed too close in the corner of a room; (ii) small obstacles, such as the leg of a chair or scattered shoes are present in the docking area, blocking part of the approach path; (iii) small obstacles such as a cable or an article of clothing are left lying in the docking area that impede the movement of the robot, for example, by causing an increased slipping of the wheels.
  • In general, the user will not intentionally cause these disturbances and will not intentionally place obstacles in the robot's path that impair access to the base station and thus a safe docking. In accordance with the embodiments described here, the attempt is made to detect such problems early and to then inform the user of them so that he may remove the disturbance. The robot, for example, is provided with a detection module that can recognize that calculating and/or travelling along the docking path has been significantly impaired or made impossible, for example, by one of the disturbances described above. The navigation sensor, for example, can be used for this by detecting obstacles in the docking area. Smaller obstacles in the docking area that the robot can drive over, for example, can be recognized by a sensor that detects an odometry error such as, for example, slipping.
  • In order to inform the user of a recognized problem, the robot possesses at least one communication interface (also known as Human Machine Interface HMI). This may comprise a visual display directly on the robot or a sound signal, in particular a voice output. In addition to this, the possibility exists of establishing a connection to an external device such as, for example, a smart phone or a tablet computer via WLAN, allowing information to be sent to the user. This information includes, for example, the nature of the detected disturbance.
  • Further, the robot can assess how severe the disturbance is and provide the user with this information (e.g. a minor problem (level I, interaction by the user not immediately needed), a relevant problem (level II, interaction by the user recommended/advisable), a grave problem (level III, interaction by the user required)). In this way the user can decide how crucial it is to intervene. For example, the base station may have been placed too close to a wall, disturbing the docking procedure but not rendering it impossible (level I or II). In this case the user can decide that the base station should remain at its present position and the robot should attempt to dock. On the other hand, the user may have accidentally placed the leg of a chair in front of the base station, rendering a direct docking maneuver impossible (level III). This problem can be quickly remedied by the user, thus ensuring the functionality of the robot. In accordance with the measures taken the user can report back to the robot via the communication interface. This report will state, for example, that; (i) the problem has been remedied; (ii) the problem is to be ignored and a docking attempted; (iii) the docking maneuver is to be postponed and the problem will be remedied later. This classification of disturbances is, naturally, only of exemplary nature and may also be carried out in any other desired manner.
  • In order that the user be informed of problems as soon as possible, the robot should always carry out the detection of disturbances immediately upon leaving the base station, hence in particular at the start of a new operational task.
  • Finally, it should be mentioned that the technical features of the devices, methods and systems described here using various examples may generally be combined with each other to arrive at other embodiments. Such combinations are generally possible and useful, unless explicitly stated here otherwise. It should be understood that all of the methods described here are to be carried out by the robot. This means that the robot includes a robot controller that is generally programmable and is, depending on the application, programmed such that the robot can carry out the respective method. The robot controller need not necessarily be realized in a single component of the robot. In general, all components that influence the externally visible behavior of the robot form a part of the robot controller. Thus, the robot controller does not have to be physically installed within the mobile robot, but may also be partially located outside of the robot in stationary (control) devices such as, e.g. computers that are connected with the robot via a communication connection.

Claims (15)

What is claimed:
1. A method for the automatic docking of an autonomous mobile robot onto a base station, the method comprising:
detecting or determining a position of the base station using a navigation module of the robot;
localizing obstacles with the navigation module of the robot, wherein the robot includes a navigation sensor;
testing whether robot access to the base station is impaired by detected obstacles.
2. The method of claim 1, wherein the testing is performed when a new task starts, the robot leaves the base station, or the position of the base station is detected.
3. The method of claim 1, further comprising informing a user, via a user interface, that the testing has revealed that robot access to the base station is impaired.
4. The method of claim 1, wherein the testing whether robot access to the base station is impaired only takes place in a certain area around the base station.
5. The method of claim 3, further comprising:
if the testing reveals that robot access to the base station is impaired, assessing the impairment caused by the obstacles in accordance with specifiable criteria; and
informing the user, via the user interface, of the assessment.
6. The method of claim 3, further comprising:
if the testing reveals that robot access to the base station is impaired, assessing the impairment caused by the obstacles based on pre-definable criteria; and
depending on the assessment, informing the user, via the user interface, of the assessment.
7. The method of claim 3, further comprising:
assessing characteristics of a floor in accordance with a specifiable criterion;
informing the user, via the user interface, that the characteristics of the floor does not meet the specifiable criterion.
8. The method of claim 3, further comprising:
providing a user interface message directly at the robot via a display, audio signal, or voice output, or at an external device connected to the robot via a communication interface.
9. A method for an autonomous mobile robot, comprising the following:
detecting geometric features of objects in an environment of the robot using a navigation module of the robot which has an electronic map;
testing whether the detected geometric features include a geometric feature that is linked with a base station; and
if one of the detected geometric features is a geometric feature linked with the base station, calculating a current position of the base station based on the geometric feature that is linked with the base station and storing the position of the base station in the electronic map.
10. The method according to claim 9, further comprising:
verifying whether the base station is actually present at positions of base stations previously stored in the map.
11. The method of claim 9, further comprising:
determining a distance between the current position of the base station and the position of the base station stored in the map;
updating the position of the base station stored in the map by the current position of the base station if the distance does not exceed a specified threshold value.
12. The method of claim 10, wherein when the verification of the position of the base station is negative, the respective previous old position in the map is deleted.
13. The method of claim 12, wherein the user is informed about changes in the map regarding the position of the base station via a user interface.
14. The method of claim 3, wherein the user is enabled to send feedback to the robot when the user has made changes in an operational area that affect base station accessibility.
15. The method of claim 1, further comprising:
performing a self-localization by the robot using as a localization hypothesis a position of the base state stored in the map at a beginning of robot deployment.
US17/536,917 2015-09-04 2021-11-29 Identification And Localization Of A Base Station Of An Autonomous Mobile Robot Abandoned US20220083070A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US17/536,917 US20220083070A1 (en) 2015-09-04 2021-11-29 Identification And Localization Of A Base Station Of An Autonomous Mobile Robot

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
DE102015114883.8 2015-09-04
DE102015114883.8A DE102015114883A1 (en) 2015-09-04 2015-09-04 Identification and localization of a base station of an autonomous mobile robot
PCT/EP2016/070761 WO2017037257A1 (en) 2015-09-04 2016-09-02 Identification and locating of a base station of an autonomous mobile robot
US201815757245A 2018-03-02 2018-03-02
US17/536,917 US20220083070A1 (en) 2015-09-04 2021-11-29 Identification And Localization Of A Base Station Of An Autonomous Mobile Robot

Related Parent Applications (2)

Application Number Title Priority Date Filing Date
PCT/EP2016/070761 Continuation WO2017037257A1 (en) 2015-09-04 2016-09-02 Identification and locating of a base station of an autonomous mobile robot
US15/757,245 Continuation US11188086B2 (en) 2015-09-04 2016-09-02 Identification and localization of a base station of an autonomous mobile robot

Publications (1)

Publication Number Publication Date
US20220083070A1 true US20220083070A1 (en) 2022-03-17

Family

ID=56883781

Family Applications (2)

Application Number Title Priority Date Filing Date
US15/757,245 Active 2037-10-29 US11188086B2 (en) 2015-09-04 2016-09-02 Identification and localization of a base station of an autonomous mobile robot
US17/536,917 Abandoned US20220083070A1 (en) 2015-09-04 2021-11-29 Identification And Localization Of A Base Station Of An Autonomous Mobile Robot

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US15/757,245 Active 2037-10-29 US11188086B2 (en) 2015-09-04 2016-09-02 Identification and localization of a base station of an autonomous mobile robot

Country Status (6)

Country Link
US (2) US11188086B2 (en)
EP (2) EP3968117A3 (en)
JP (2) JP6971223B2 (en)
CN (2) CN114740852A (en)
DE (1) DE102015114883A1 (en)
WO (1) WO2017037257A1 (en)

Families Citing this family (37)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9259838B1 (en) * 2014-07-24 2016-02-16 Google Inc. Systems and methods for ground plane estimation
JP6916812B2 (en) * 2016-04-14 2021-08-11 デカ・プロダクツ・リミテッド・パートナーシップ User control device for transporter
US10016896B2 (en) * 2016-06-30 2018-07-10 Brain Corporation Systems and methods for robotic behavior around moving bodies
CN106308685B (en) * 2016-08-23 2019-10-11 北京小米移动软件有限公司 cleaning robot and control method thereof
KR102461938B1 (en) * 2017-04-04 2022-10-31 엘지전자 주식회사 Method of identifying obstacle on a driving surface and robot implementing thereof
CN106980320B (en) * 2017-05-18 2020-06-19 上海思岚科技有限公司 Robot charging method and device
EP3428759B1 (en) * 2017-07-13 2020-09-02 Vorwerk & Co. Interholding GmbH Method for operating an automatically moving service device
DE102018116065A1 (en) * 2017-07-13 2019-01-17 Vorwerk & Co. Interholding Gmbh Method for operating a self-propelled service device
US10383499B2 (en) * 2017-08-31 2019-08-20 Irobot Corporation Wet robot docking station
CN114504276B (en) * 2017-09-25 2023-11-28 北京石头创新科技有限公司 Autonomous mobile robot, pile searching method thereof and control device
US10365656B2 (en) * 2017-11-22 2019-07-30 Locus Robotics Corp. Robot charger docking localization
KR102203439B1 (en) * 2018-01-17 2021-01-14 엘지전자 주식회사 a Moving robot and Controlling method for the moving robot
EP3770712B1 (en) * 2018-03-19 2022-09-21 Honda Motor Co., Ltd. Autonomous traveling work machine
EP3778145B1 (en) * 2018-04-06 2024-05-29 LG Electronics Inc. Mobile robot and control method of mobile robot
CN110389341B (en) * 2018-04-18 2021-06-25 深圳市优必选科技有限公司 Charging pile identification method and device, robot and computer readable storage medium
DE102018206067A1 (en) * 2018-04-20 2019-10-24 Robert Bosch Gmbh Method and device for determining a highly accurate position of a vehicle
CN116509280A (en) 2018-07-19 2023-08-01 科沃斯机器人股份有限公司 Robot control method, robot, and storage medium
US11009887B2 (en) * 2018-07-26 2021-05-18 Toyota Research Institute, Inc. Systems and methods for remote visual inspection of a closed space
CN109683605B (en) 2018-09-25 2020-11-24 上海肇观电子科技有限公司 Robot and automatic recharging method and system thereof, electronic equipment and storage medium
KR102220865B1 (en) * 2018-10-02 2021-02-25 엘지전자 주식회사 Method of adaptive operation in unevenness height of space and device implementing thereof
CN109586360B (en) * 2018-11-09 2020-09-22 深圳市银星智能科技股份有限公司 Robot automatic charging method and device, charging pile and robot
WO2020117766A1 (en) * 2018-12-03 2020-06-11 Sharkninja Operating Llc Optical indicium for communicating information to autonomous devices
DE102019202299B4 (en) * 2019-02-20 2020-12-31 Zf Friedrichshafen Ag On-line calibration and calibration setup procedures
CN111743450B (en) * 2019-03-26 2021-06-08 速感科技(北京)有限公司 Recharging method for movable device and movable device
US11493336B2 (en) * 2020-06-22 2022-11-08 Pixart Imaging Inc. Optical navigation device which can determine dirtiness level of cover or fix multi light pattern issue
CN110622085A (en) * 2019-08-14 2019-12-27 珊口(深圳)智能科技有限公司 Mobile robot and control method and control system thereof
CN110597249B (en) 2019-08-23 2022-08-05 深圳市优必选科技股份有限公司 Robot and recharging positioning method and device thereof
KR102275083B1 (en) * 2019-12-05 2021-07-08 주식회사 제타뱅크 Robotic systems and a returning method of robot for automatic charging
CN111813119B (en) * 2020-07-11 2024-02-27 广东翼景信息科技有限公司 High-precision navigation method for inspection robot
EP3968051A1 (en) * 2020-09-15 2022-03-16 Infineon Technologies AG Guiding system for a robot, base station including such a guiding system, and method for guiding a robot
KR20220046190A (en) * 2020-10-07 2022-04-14 삼성전자주식회사 Robot, charghing station and robot charghing system comprising the same
CN112346453A (en) * 2020-10-14 2021-02-09 深圳市杉川机器人有限公司 Automatic robot recharging method and device, robot and storage medium
US11826906B2 (en) * 2020-11-24 2023-11-28 Pixart Imaging Inc. Method for eliminating misjudgment of reflective light and optical sensing system
US11921205B2 (en) 2020-11-24 2024-03-05 Pixart Imaging Inc. Method for eliminating misjudgment of reflective lights and optical sensing system
US11969140B2 (en) * 2021-06-22 2024-04-30 Micron Technology, Inc. Surface cleaning
CN115890647A (en) * 2021-08-27 2023-04-04 成都家有为力机器人技术有限公司 Robot base positioning method and device based on visual navigation, robot and medium
CN113892857B (en) * 2021-10-29 2023-04-28 珠海格力电器股份有限公司 Cleaning robot cleaning control method, device, computer equipment and medium

Citations (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE4340771A1 (en) * 1993-06-08 1994-12-15 Samsung Electronics Co Ltd Automatic cleaning device
CN1299083A (en) * 1999-10-29 2001-06-13 索尼公司 Robot and charging system, charging station searching method, connector and electric connection structure
JP2004283956A (en) * 2003-03-20 2004-10-14 Sony Corp Charging system, robot device, charging device, and charging method
US20060238159A1 (en) * 2005-04-25 2006-10-26 Lg Electronics Inc. Position calculation system for mobile robot and charging-stand return and method using the same
JP2009101044A (en) * 2007-10-25 2009-05-14 Panasonic Corp Self-propelled apparatus
US20110130874A1 (en) * 2009-11-27 2011-06-02 Chuan-Ching Tsao Moving devices and controlling methods therefor
WO2014043732A1 (en) * 2012-09-24 2014-03-27 RobArt GmbH Robot and method for autonomous inspection or processing of floor areas
US20140278252A1 (en) * 2013-03-14 2014-09-18 Ecolab Usa Inc. System and method for monitoring of floor conditions
KR101484941B1 (en) * 2009-06-12 2015-01-23 삼성전자주식회사 Robot cleaner and control method thereof
JP2015534048A (en) * 2012-08-27 2015-11-26 アクティエボラゲット エレクトロラックス Robot positioning system
US20160147406A1 (en) * 2014-11-25 2016-05-26 Samsung Electronics Co., Ltd. Method for providing graphical user interface and electronic device for supporting the same
CN105629971A (en) * 2014-11-03 2016-06-01 贵州亿丰升华科技机器人有限公司 Robot automatic charging system and control method therefor
CN105700522A (en) * 2014-11-11 2016-06-22 沈阳新松机器人自动化股份有限公司 Robot charging method and charging system thereof
KR20160097966A (en) * 2015-02-10 2016-08-18 에브리봇 주식회사 A robot cleaner and a method for operating it
EP2749194B1 (en) * 2012-12-26 2018-11-14 LG Electronics, Inc. Automatic cleaner
DE112017003497T5 (en) * 2016-07-11 2019-05-02 Groove X, Inc. Self-acting robot with controlled amount of activity

Family Cites Families (265)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4674048A (en) 1983-10-26 1987-06-16 Automax Kabushiki-Kaisha Multiple robot control system using grid coordinate system for tracking and completing travel over a mapped region containing obstructions
JPS61251809A (en) 1985-05-01 1986-11-08 Hitachi Ltd Automatic focus adjusting device
US4777416A (en) 1986-05-16 1988-10-11 Denning Mobile Robotics, Inc. Recharge docking system for mobile robot
US5377106A (en) 1987-03-24 1994-12-27 Fraunhofer Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Process for navigating an unmanned vehicle and a vehicle for the same
JPH0313611A (en) 1989-06-07 1991-01-22 Toshiba Corp Automatic cleaner
US5109566A (en) 1990-06-28 1992-05-05 Matsushita Electric Industrial Co., Ltd. Self-running cleaning apparatus
US5260710A (en) 1991-01-31 1993-11-09 Stanley Electric Co., Ltd. Vehicular optical-radar apparatus
JP3198532B2 (en) 1991-05-14 2001-08-13 松下電器産業株式会社 Self-propelled vacuum cleaner
JPH0680203A (en) 1992-03-24 1994-03-22 East Japan Railway Co Control method for floor surface cleaning robot
CA2115859C (en) 1994-02-23 1995-12-26 Brian Dewan Method and apparatus for optimizing sub-pixel resolution in a triangulation based distance measuring device
DE4421805C1 (en) 1994-06-22 1995-08-17 Siemens Ag Orientation motion and control of autonomous mobile robot
JP3346513B2 (en) 1994-07-01 2002-11-18 ミノルタ株式会社 Map storage method and route creation method using the map
BE1008470A3 (en) 1994-07-04 1996-05-07 Colens Andre Device and automatic system and equipment dedusting sol y adapted.
US5995884A (en) 1997-03-07 1999-11-30 Allen; Timothy P. Computer peripheral floor cleaning system and navigation method
FR2763726B1 (en) 1997-05-20 2003-01-17 Bouchaib Hoummadi METHOD FOR MANAGING ROAD TRAFFIC BY VIDEO CAMERA
US6532404B2 (en) 1997-11-27 2003-03-11 Colens Andre Mobile robots and their control system
ES2172936T3 (en) 1997-11-27 2002-10-01 Solar & Robotics IMPROVEMENTS IN MOBILE ROBOTS AND THEIR CONTROL SYSTEM.
KR20010041694A (en) 1998-03-10 2001-05-25 칼 하인쯔 호르닝어 Optical sensor system for detecting the position of an object
IL124413A (en) 1998-05-11 2001-05-20 Friendly Robotics Ltd System and method for area coverage with an autonomous robot
AU4999899A (en) 1998-07-20 2000-02-07 Procter & Gamble Company, The Robotic system
WO2000078410A1 (en) 1999-06-17 2000-12-28 Solar & Robotics S.A. Device for automatically picking up objects
GB9917232D0 (en) 1999-07-23 1999-09-22 Notetry Ltd Method of operating a floor cleaning device
US7155308B2 (en) 2000-01-24 2006-12-26 Irobot Corporation Robot obstacle detection system
US6594844B2 (en) 2000-01-24 2003-07-22 Irobot Corporation Robot obstacle detection system
JP2002085305A (en) 2000-09-12 2002-03-26 Toshiba Tec Corp Robot cleaner and robot cleaner system
KR20020076297A (en) 2000-12-11 2002-10-09 미쓰비시덴키 가부시키가이샤 Optical distance sensor
US6690134B1 (en) 2001-01-24 2004-02-10 Irobot Corporation Method and system for robot localization and confinement
US7571511B2 (en) 2002-01-03 2009-08-11 Irobot Corporation Autonomous floor-cleaning robot
US6883201B2 (en) 2002-01-03 2005-04-26 Irobot Corporation Autonomous floor-cleaning robot
JP3594016B2 (en) 2001-01-30 2004-11-24 日本電気株式会社 Robot program execution method, robot system and program processing device
RU2220643C2 (en) 2001-04-18 2004-01-10 Самсунг Гванджу Электроникс Ко., Лтд. Automatic cleaning apparatus, automatic cleaning system and method for controlling of system (versions)
US20040187457A1 (en) 2001-05-28 2004-09-30 Andre Colens Robotic lawnmower
US7663333B2 (en) 2001-06-12 2010-02-16 Irobot Corporation Method and system for multi-mode coverage for an autonomous robot
ES2660836T3 (en) 2001-06-12 2018-03-26 Irobot Corporation Multi-code coverage method and system for an autonomous robot
US6667592B2 (en) 2001-08-13 2003-12-23 Intellibot, L.L.C. Mapped robot system
EP1441632B1 (en) 2001-09-26 2013-05-01 F. Robotics Acquisitions Ltd. Robotic vacuum cleaner
DE10204223B4 (en) 2002-01-31 2004-05-06 Infineon Technologies Ag Housing for a coupling arrangement for coupling in and / or coupling out optical signals
JP2004133882A (en) * 2002-05-10 2004-04-30 Royal Appliance Mfg Co Autonomous multi-platform robot system
JP2003330543A (en) 2002-05-17 2003-11-21 Toshiba Tec Corp Charging type autonomous moving system
AU2002335204A1 (en) 2002-10-04 2004-04-23 Fujitsu Limited Robot system and autonomously traveling robot
DE10262191A1 (en) 2002-12-23 2006-12-14 Alfred Kärcher Gmbh & Co. Kg Mobile tillage implement
KR100561855B1 (en) 2002-12-30 2006-03-16 삼성전자주식회사 Robot localization system
US7805220B2 (en) 2003-03-14 2010-09-28 Sharper Image Acquisition Llc Robot vacuum with internal mapping system
US20050010331A1 (en) 2003-03-14 2005-01-13 Taylor Charles E. Robot vacuum with floor type modes
US7756322B2 (en) 2003-08-18 2010-07-13 Honda Motor Co., Ltd. Picture taking mobile robot
US20070061041A1 (en) 2003-09-02 2007-03-15 Zweig Stephen E Mobile robot with wireless location sensing apparatus
EP1533629A3 (en) 2003-11-21 2006-05-24 Siemens Aktiengesellschaft Distance measurement with a mobile terminal
US7332890B2 (en) 2004-01-21 2008-02-19 Irobot Corporation Autonomous robot auto-docking and energy management systems and methods
DE102004004505B9 (en) 2004-01-22 2010-08-05 Alfred Kärcher Gmbh & Co. Kg Soil cultivation device and method for its control
JP4264009B2 (en) 2004-01-23 2009-05-13 シャープ株式会社 Self-propelled vacuum cleaner
AU2004316426B2 (en) 2004-01-28 2010-08-19 Irobot Corporation Debris sensor for cleaning apparatus
JP3841220B2 (en) 2004-01-30 2006-11-01 船井電機株式会社 Autonomous traveling robot cleaner
JP2005211359A (en) 2004-01-30 2005-08-11 Funai Electric Co Ltd Autonomous traveling robot cleaner system
US7729801B2 (en) 2004-02-03 2010-06-01 F Robotics Acquisitions Ltd. Robot docking station and robot for use therewith
KR100506097B1 (en) 2004-02-04 2005-08-03 삼성전자주식회사 Method and apparatus for making magnetic field map and method and apparatus for checking pose of the moving body using the map
US20060020369A1 (en) 2004-03-11 2006-01-26 Taylor Charles E Robot vacuum cleaner
JP2005270413A (en) 2004-03-25 2005-10-06 Funai Electric Co Ltd Self-propelled vacuum cleaner
JP2007530978A (en) 2004-03-29 2007-11-01 エヴォリューション ロボティクス インコーポレイテッド Position estimation method and apparatus using reflected light source
JP4436186B2 (en) * 2004-05-12 2010-03-24 アルパイン株式会社 Navigation device and map display method
JP4377744B2 (en) 2004-05-13 2009-12-02 本田技研工業株式会社 Robot controller
SG174000A1 (en) 2004-06-24 2011-09-29 Irobot Corp Remote control scheduler and method for autonomous robotic device
US8972052B2 (en) 2004-07-07 2015-03-03 Irobot Corporation Celestial navigation system for an autonomous vehicle
KR100641113B1 (en) 2004-07-30 2006-11-02 엘지전자 주식회사 Mobile robot and his moving control method
KR100645379B1 (en) 2004-10-29 2006-11-15 삼성광주전자 주식회사 A robot controlling system and a robot control method
US7620476B2 (en) 2005-02-18 2009-11-17 Irobot Corporation Autonomous surface cleaning robot for dry cleaning
US7389156B2 (en) 2005-02-18 2008-06-17 Irobot Corporation Autonomous surface cleaning robot for wet and dry cleaning
EP1850725B1 (en) 2005-02-18 2010-05-19 iRobot Corporation Autonomous surface cleaning robot for wet and dry cleaning
KR100638220B1 (en) 2005-04-23 2006-10-27 엘지전자 주식회사 Position sensing device of mobile robot and robot cleaner equipped with it
JP4455417B2 (en) 2005-06-13 2010-04-21 株式会社東芝 Mobile robot, program, and robot control method
DE202005021588U1 (en) 2005-09-05 2008-10-09 Robert Bosch Gmbh Laser rangefinder
ATE524784T1 (en) 2005-09-30 2011-09-15 Irobot Corp COMPANION ROBOTS FOR PERSONAL INTERACTION
US9002511B1 (en) 2005-10-21 2015-04-07 Irobot Corporation Methods and systems for obstacle detection using structured light
ES2623920T3 (en) 2005-12-02 2017-07-12 Irobot Corporation Robot system
KR101300492B1 (en) 2005-12-02 2013-09-02 아이로보트 코퍼레이션 Coverage robot mobility
WO2008013568A2 (en) 2005-12-30 2008-01-31 Irobot Corporation Autonomous mobile robot
DE102006007764A1 (en) 2006-02-20 2007-08-23 Sick Ag Optoelectronic device and method for its operation
KR100988736B1 (en) 2006-03-15 2010-10-20 삼성전자주식회사 Home network system and method for moving the shortest path of autonomous mobile robot
US7483151B2 (en) 2006-03-17 2009-01-27 Alpineon D.O.O. Active 3D triangulation-based imaging method and device
US7861366B2 (en) 2006-04-04 2011-01-04 Samsung Electronics Co., Ltd. Robot cleaner system having robot cleaner and docking station
KR100735565B1 (en) 2006-05-17 2007-07-04 삼성전자주식회사 Method for detecting an object using structured light and robot using the same
KR100791382B1 (en) 2006-06-01 2008-01-07 삼성전자주식회사 Method for classifying and collecting of area features as robot's moving path and robot controlled as the area features, apparatus and method for composing user interface using area features
US8355818B2 (en) 2009-09-03 2013-01-15 Battelle Energy Alliance, Llc Robots, systems, and methods for hazard evaluation and visualization
KR100791384B1 (en) 2006-07-05 2008-01-07 삼성전자주식회사 Method for dividing regions by feature points and apparatus thereof and mobile cleaning robot
KR100791386B1 (en) 2006-08-18 2008-01-07 삼성전자주식회사 Method and system of cell decomposition in mobile robot
US8996172B2 (en) 2006-09-01 2015-03-31 Neato Robotics, Inc. Distance sensor system and method
KR100815545B1 (en) 2006-12-06 2008-03-20 삼성광주전자 주식회사 The method of charging service robot
KR100791389B1 (en) 2006-12-26 2008-01-07 삼성전자주식회사 Apparatus and method for measuring distance using structured light
DE102007003024A1 (en) 2007-01-20 2008-07-31 Sick Ag Triangulation sensor with distance determination from light spot position and shape
DE102007016913A1 (en) 2007-04-05 2008-10-09 Inmach Intelligente Maschinen Gmbh Method for running a work surface
DE102007016802B3 (en) 2007-04-05 2008-05-15 Miele & Cie. Kg Self-propelled tilling device e.g. robot, navigating method, involves determining driving direction by evaluating determined vectors and by controlling distance to preceding tracks, and forming meander/spiral shaped preset track
EP2574264B1 (en) 2007-05-09 2016-03-02 iRobot Corporation Compact autonomous coverage robot
US20090048727A1 (en) 2007-08-17 2009-02-19 Samsung Electronics Co., Ltd. Robot cleaner and control method and medium of the same
KR20090019338A (en) 2007-08-20 2009-02-25 삼성전자주식회사 Optical sensor
KR101330734B1 (en) 2007-08-24 2013-11-20 삼성전자주식회사 Robot cleaner system having robot cleaner and docking station
US20090062974A1 (en) * 2007-09-03 2009-03-05 Junichi Tamamoto Autonomous Mobile Robot System
DE202007014849U1 (en) 2007-10-24 2008-01-17 Pepperl + Fuchs Gmbh Optoelectronic sensor for detecting objects in a surveillance area
KR101415879B1 (en) 2008-01-04 2014-07-07 삼성전자 주식회사 Method and apparatus for docking moving robot
KR20090077547A (en) 2008-01-11 2009-07-15 삼성전자주식회사 Method and apparatus of path planning for a mobile robot
WO2009090807A1 (en) * 2008-01-16 2009-07-23 Nec Corporation Mobile device, method for moving mobile device, and program for controlling movement of mobile device
DE102008014912B4 (en) 2008-03-19 2023-01-19 Vorwerk & Co. Interholding Gmbh Automatically movable floor dust collector
JP4909930B2 (en) 2008-03-28 2012-04-04 日立アプライアンス株式会社 Self-propelled cleaning system
US8194233B2 (en) 2008-04-11 2012-06-05 Microsoft Corporation Method and system to reduce stray light reflection error in time-of-flight sensor arrays
EP2303085B1 (en) 2008-04-24 2017-07-26 iRobot Corporation Application of localization, positioning & navigation systems for robotic enabled mobile products
DE102008028931A1 (en) 2008-06-18 2009-12-24 BSH Bosch und Siemens Hausgeräte GmbH Robot i.e. dust collecting robot, drive movement controlling method, involves stopping drive movement of robot during determination of coordinate values based on comparison of coordinate values of virtual partial region and/or virtual wall
CN101387514B (en) 2008-08-28 2010-07-28 上海科勒电子科技有限公司 Distance detecting induction device
KR101553654B1 (en) 2009-02-13 2015-10-01 삼성전자 주식회사 Mobile robot and method for moving of mobile robot
DE102009001734A1 (en) 2009-03-23 2011-02-24 Robert Bosch Gmbh optics carrier
JP5506225B2 (en) 2009-03-30 2014-05-28 セーレン株式会社 How to recycle advertising sheets
CN101923351B (en) 2009-06-12 2015-03-04 三星电子株式会社 Robot cleaner and control method thereof
US8428776B2 (en) * 2009-06-18 2013-04-23 Michael Todd Letsky Method for establishing a desired area of confinement for an autonomous robot and autonomous robot implementing a control system for executing the same
KR101672787B1 (en) 2009-06-19 2016-11-17 삼성전자주식회사 Robot cleaner and docking station and robot cleaner system having the same and control method thereof
DE102009059214A1 (en) * 2009-08-12 2011-02-17 Vorwerk & Co. Interholding Gmbh Automatically movable device and method for the route guidance of such a device
DE102009041362A1 (en) 2009-09-11 2011-03-24 Vorwerk & Co. Interholding Gmbh Method for operating a cleaning robot
CN104970741B (en) 2009-11-06 2017-08-29 艾罗伯特公司 Method and system for surface to be completely covered by autonomous humanoid robot
DE102009052629A1 (en) 2009-11-10 2011-05-12 Vorwerk & Co. Interholding Gmbh Method for controlling a robot
KR101626984B1 (en) 2009-11-16 2016-06-02 엘지전자 주식회사 Robot cleaner and controlling method of the same
JP2011128899A (en) 2009-12-17 2011-06-30 Murata Machinery Ltd Autonomous mobile device
US8892251B1 (en) 2010-01-06 2014-11-18 Irobot Corporation System and method for autonomous mopping of a floor surface
DE102010000174B4 (en) 2010-01-22 2022-09-01 Vorwerk & Co. Interholding Gesellschaft mit beschränkter Haftung Method for cleaning a room using an automatically movable cleaning device
KR101686170B1 (en) 2010-02-05 2016-12-13 삼성전자주식회사 Apparatus for planning traveling path and method thereof
DE102010000317A1 (en) 2010-02-05 2011-08-11 Vorwerk & Co. Interholding GmbH, 42275 Method for cleaning room by using automatically movable cleaning device, involves registering rectangle to be projected in room with longest possible dimensions by using algorithm
KR101649645B1 (en) 2010-02-08 2016-08-22 엘지전자 주식회사 Robot cleaner and controlling method thereof
JP2011181997A (en) 2010-02-26 2011-09-15 Brother Industries Ltd Communication device and program
DE102010000607B4 (en) * 2010-03-02 2022-06-15 Vorwerk & Co. Interholding Gmbh Household vacuum cleaner that can be used as a base station for an automatically movable suction and/or sweeping device
KR20110119118A (en) 2010-04-26 2011-11-02 엘지전자 주식회사 Robot cleaner, and remote monitoring system using the same
KR101487778B1 (en) 2010-05-11 2015-01-29 삼성전자 주식회사 Sensing system and moving robot having the same
US8442682B2 (en) * 2010-05-28 2013-05-14 Toyota Motor Engineering & Manufacturing North America, Inc. Autonomous robot charging stations and methods
DE102010017211A1 (en) 2010-06-02 2011-12-08 Vorwerk & Co. Interholding Gmbh Method for cleaning floor e.g. hard floor in household area, involves holding cleaning tool and/or cleaning agent or cleaning fluid in base station via floor cleaning device for cleaning different regions of floor, after recognizing stain
DE102010017689A1 (en) 2010-07-01 2012-01-05 Vorwerk & Co. Interholding Gmbh Automatically movable device and method for orientation of such a device
JP5560978B2 (en) 2010-07-13 2014-07-30 村田機械株式会社 Autonomous mobile
KR101483541B1 (en) 2010-07-15 2015-01-19 삼성전자주식회사 Autonomous cleaning device, maintenance station and cleaning system having them
DE102010033768A1 (en) 2010-08-09 2012-02-09 Dürr Systems GmbH Control system and control method for a robot
DE102011050357A1 (en) 2010-08-12 2012-02-16 Vorwerk & Co. Interholding Gmbh Method for controlling i.e. guiding, movable household floor cleaning device e.g. sucking robot, involves recording and processing light signal such that transmission of light signal to cleaning area is controlled
CN101945325B (en) 2010-08-13 2016-04-27 厦门雅迅网络股份有限公司 A kind of friend's cognitive method based on architecture
CN102407522B (en) 2010-09-19 2014-03-26 泰怡凯电器(苏州)有限公司 Intelligent robot system and charging butting method thereof
KR20120043865A (en) 2010-10-27 2012-05-07 주식회사 케이티 System, method and apparatus for providing robot interaction services using location information of mobile communication terminal
KR101750340B1 (en) 2010-11-03 2017-06-26 엘지전자 주식회사 Robot cleaner and controlling method of the same
KR101752190B1 (en) 2010-11-24 2017-06-30 삼성전자주식회사 Robot cleaner and method for controlling the same
EP2659260B1 (en) 2010-12-30 2019-11-20 iRobot Corporation Debris monitoring
WO2012099694A1 (en) 2010-12-30 2012-07-26 Irobot Corporation Coverage robot navigation
US9749780B2 (en) 2011-02-05 2017-08-29 Apple Inc. Method and apparatus for mobile location determination
US20120215380A1 (en) 2011-02-23 2012-08-23 Microsoft Corporation Semi-autonomous robot that supports multiple modes of navigation
US8779391B2 (en) 2011-03-03 2014-07-15 Teckni-Corp Sterilization system with ultraviolet emitter for eradicating biological contaminants
DE102011006062B4 (en) 2011-03-24 2023-05-25 RobArt GmbH Procedure for autonomous inspection of an environment or processing of ground surfaces
KR101842460B1 (en) 2011-04-12 2018-03-27 엘지전자 주식회사 Robot cleaner, and remote monitoring system and method of the same
KR101850386B1 (en) 2011-04-19 2018-04-19 엘지전자 주식회사 Robot cleaner and controlling method of the same
EP3878333B1 (en) 2011-04-29 2023-12-13 iRobot Corporation An autonomous mobile robot for cleaning with a front roller in a first horizontal plane positioned above a second horizontal plane of a rear roller
KR101760950B1 (en) 2011-05-17 2017-07-24 엘지전자 주식회사 Controlling mehtod of network system
JP2013007708A (en) 2011-06-27 2013-01-10 Hitachi High-Technologies Corp Inspection apparatus and method for concealing inspection recipe
JP5399525B2 (en) 2011-06-29 2014-01-29 シャープ株式会社 Optical distance measuring device and electronic device
DE102011051729A1 (en) 2011-07-11 2013-01-17 Alfred Kärcher Gmbh & Co. Kg Self-propelled floor cleaning device
TW201305761A (en) 2011-07-21 2013-02-01 Ememe Robot Co Ltd An autonomous robot and a positioning method thereof
US8761933B2 (en) 2011-08-02 2014-06-24 Microsoft Corporation Finding a called party
KR101366860B1 (en) 2011-09-20 2014-02-21 엘지전자 주식회사 Mobile robot and controlling method of the same
JP5348215B2 (en) * 2011-09-29 2013-11-20 カシオ計算機株式会社 Information acquisition apparatus, information acquisition method, information acquisition program, and information acquisition system
US8798840B2 (en) 2011-09-30 2014-08-05 Irobot Corporation Adaptive mapping with spatial summaries of sensor data
JP6087498B2 (en) * 2011-10-06 2017-03-01 日本信号株式会社 Self-propelled device charging system
WO2013085085A1 (en) 2011-12-08 2013-06-13 Lg Electronics Inc. Automatic moving apparatus and manual operation method thereof
JP2013146302A (en) 2012-01-17 2013-08-01 Sharp Corp Self-propelled electronic device
US8982217B1 (en) 2012-01-31 2015-03-17 Google Inc. Determining states and modifying environments according to states
DE102012201870A1 (en) * 2012-02-08 2013-08-08 RobArt GmbH Method for automatically triggering a self-localization
KR101984214B1 (en) 2012-02-09 2019-05-30 삼성전자주식회사 Apparatus and method for controlling cleaning in rototic cleaner
JP5777545B2 (en) * 2012-03-09 2015-09-09 三菱電機株式会社 Electric vacuum cleaner
US9146560B2 (en) * 2012-03-30 2015-09-29 Irobot Corporation System and method for implementing force field deterrent for robot
CN102738862B (en) 2012-06-13 2014-12-03 杭州瓦瑞科技有限公司 Automatic charging system for movable robot
DE102012105608A1 (en) 2012-06-27 2014-01-02 Miele & Cie. Kg Self-propelled cleaning device and method for operating a self-propelled cleaning device
DE102012211071B3 (en) 2012-06-27 2013-11-21 RobArt GmbH Interaction between a mobile robot and an alarm system
WO2014017256A1 (en) 2012-07-24 2014-01-30 シャープ株式会社 Optical distance measuring apparatus and electronic apparatus
US8855914B1 (en) 2012-08-31 2014-10-07 Neato Robotics, Inc. Method and apparatus for traversing corners of a floored area with a robotic surface treatment apparatus
JP5735940B2 (en) * 2012-09-11 2015-06-17 シャープ株式会社 Server, control system, self-propelled cleaner, program, and recording medium
TWI459170B (en) 2012-10-04 2014-11-01 Ind Tech Res Inst A moving control device and an automatic guided vehicle with the same
US9538892B2 (en) 2012-10-05 2017-01-10 Irobot Corporation Robot management systems for determining docking station pose including mobile robots and methods using same
US8972061B2 (en) 2012-11-02 2015-03-03 Irobot Corporation Autonomous coverage robot
US20140128093A1 (en) 2012-11-06 2014-05-08 Qualcomm Incorporated Portal transition parameters for use in mobile device positioning
TWI481980B (en) 2012-12-05 2015-04-21 Univ Nat Chiao Tung Electronic apparatus and navigation method thereof
KR101428877B1 (en) 2012-12-05 2014-08-14 엘지전자 주식회사 A robot cleaner
KR20140073854A (en) 2012-12-07 2014-06-17 주식회사 유진로봇 Obstacle detect system using psd scanner in vaccum robot
DE102012112036B4 (en) 2012-12-10 2023-06-07 Miele & Cie. Kg Self-propelled tillage implement and method for navigating a self-propelled tillage implement
DE102012112035A1 (en) 2012-12-10 2014-06-12 Miele & Cie. Kg Robot vacuum cleaner operating method, involves transferring processing date to vacuum cleaner in operation of robot vacuum cleaner as result of processing control information returns to cleaner for activation of cleaner by individual
KR102058918B1 (en) 2012-12-14 2019-12-26 삼성전자주식회사 Home monitoring method and apparatus
US9903130B2 (en) * 2012-12-22 2018-02-27 Maytronics Ltd. Autonomous pool cleaning robot with an external docking station
DE102013100192A1 (en) 2013-01-10 2014-07-10 Miele & Cie. Kg Self-propelled robot and method for distance determination in a self-propelled robot
US9233472B2 (en) 2013-01-18 2016-01-12 Irobot Corporation Mobile robot providing environmental mapping for household environmental control
US9375847B2 (en) 2013-01-18 2016-06-28 Irobot Corporation Environmental management systems including mobile robots and methods using same
WO2014113091A1 (en) 2013-01-18 2014-07-24 Irobot Corporation Environmental management systems including mobile robots and methods using same
GB2509991B (en) 2013-01-22 2015-03-11 Dyson Technology Ltd Docking station for a mobile robot
GB2509989B (en) 2013-01-22 2015-03-04 Dyson Technology Ltd Docking station for a mobile robot
GB2509990B (en) 2013-01-22 2014-12-10 Dyson Technology Ltd Docking station for a mobile robot
KR101450537B1 (en) 2013-03-05 2014-10-14 엘지전자 주식회사 Robot cleaner
KR101490170B1 (en) 2013-03-05 2015-02-05 엘지전자 주식회사 Robot cleaner
KR101450569B1 (en) 2013-03-05 2014-10-14 엘지전자 주식회사 Robot cleaner
WO2014138472A2 (en) 2013-03-06 2014-09-12 Robotex Inc. System and method for collecting and processing data and for utilizing robotic and/or human resources
JP6088858B2 (en) 2013-03-12 2017-03-01 シャープ株式会社 Self-propelled equipment
JP2014197294A (en) 2013-03-29 2014-10-16 株式会社日立産機システム Position identification device and mobile robot having the same
JP2014200449A (en) * 2013-04-04 2014-10-27 シャープ株式会社 Self-propelled vacuum cleaner
KR102071575B1 (en) 2013-04-23 2020-01-30 삼성전자 주식회사 Moving robot, user terminal apparatus, and control method thereof
CN104117987B (en) 2013-04-26 2017-05-10 恩斯迈电子(深圳)有限公司 Mobile robot
DE102013104399A1 (en) 2013-04-30 2014-10-30 Vorwerk & Co. Interholding Gmbh Base station for a self-moving device
DE102013104547A1 (en) 2013-05-03 2014-11-06 Miele & Cie. Kg Self-propelled tillage implement and optical system for a self-propelled tillage implement
GB2513912B (en) 2013-05-10 2018-01-24 Dyson Technology Ltd Apparatus for guiding an autonomous vehicle towards a docking station
CN104161487B (en) 2013-05-17 2018-09-04 恩斯迈电子(深圳)有限公司 Mobile device
KR20140145648A (en) 2013-06-13 2014-12-24 삼성전자주식회사 Cleaning robot and method for controlling the same
KR101534005B1 (en) 2013-07-16 2015-07-06 주식회사 유진로봇 System for cleaning user defined area using cleaning robot and method thereof
JP6406680B2 (en) * 2013-08-06 2018-10-17 アルフレッド ケルヒャー エスエー ウント コンパニー カーゲー Method for operating a floor cleaning device and floor cleaning device
JP5897517B2 (en) 2013-08-21 2016-03-30 シャープ株式会社 Autonomous mobile
CN104460663A (en) 2013-09-23 2015-03-25 科沃斯机器人科技(苏州)有限公司 Method for controlling cleaning robot through smart phone
US20150115138A1 (en) 2013-10-25 2015-04-30 Avago Technologies General Ip (Singapore) Pte. Ltd. Sensing Device With A Shield
KR102095817B1 (en) 2013-10-31 2020-04-01 엘지전자 주식회사 Mobile robot, charging apparatus for the mobile robot, and mobile robot system
KR102152641B1 (en) 2013-10-31 2020-09-08 엘지전자 주식회사 Mobile robot
EP2870852A1 (en) 2013-11-11 2015-05-13 Honda Research Institute Europe GmbH Lawn mower with network communication
US9573275B2 (en) 2013-11-12 2017-02-21 Husqvarna Ab Navigation for a robotic working tool
KR20160095029A (en) 2013-12-06 2016-08-10 알프레드 캐르혀 게엠베하 운트 컴파니. 카게 Cleaning system
US10045675B2 (en) 2013-12-19 2018-08-14 Aktiebolaget Electrolux Robotic vacuum cleaner with side brush moving in spiral pattern
KR102118049B1 (en) 2013-12-19 2020-06-09 엘지전자 주식회사 robot cleaner, robot cleaner system and a control method of the same
JP6750921B2 (en) * 2013-12-19 2020-09-02 アクチエボラゲット エレクトロルックス Robot vacuum cleaner
CN203672362U (en) 2013-12-31 2014-06-25 科沃斯机器人科技(苏州)有限公司 Laser distance measurement device and self-moving robot with same
KR102118051B1 (en) 2014-01-17 2020-06-02 엘지전자 주식회사 robot cleaning system and charge method of the same
KR102158695B1 (en) 2014-02-12 2020-10-23 엘지전자 주식회사 robot cleaner and a control method of the same
JP5543042B1 (en) * 2014-02-12 2014-07-09 株式会社コスモライフ pedestal
KR102072387B1 (en) 2014-03-20 2020-02-03 삼성전자주식회사 Robot cleaner and method for controlling the same
DE102014012811B4 (en) 2014-03-27 2017-09-21 Miele & Cie. Kg Floor cleaning apparatus and method and system for determining a floor plan by a self-propelled floor cleaning device
CN104972462B (en) 2014-04-14 2017-04-19 科沃斯机器人股份有限公司 Obstacle avoidance walking method of self-moving robot
KR101578864B1 (en) 2014-04-25 2015-12-29 에브리봇 주식회사 Distance sensor, robot cleaner and control method thereof
KR101578884B1 (en) 2014-04-25 2015-12-29 에브리봇 주식회사 Distance sensor, robot cleaner and control method thereof
KR101578861B1 (en) 2014-04-25 2015-12-21 에브리봇 주식회사 Distance sensor, robot cleaner and control method thereof
KR101578878B1 (en) 2014-05-07 2015-12-21 에브리봇 주식회사 Distance sensor device, robot cleaner and control method thereof
JP5676039B1 (en) 2014-05-28 2015-02-25 シャープ株式会社 Self-propelled device, control method for self-propelled device, and control program for self-propelled device
CN105334847B (en) 2014-06-26 2021-07-27 科沃斯机器人股份有限公司 Self-moving robot
DE102014110104B4 (en) 2014-07-18 2016-12-15 eyeworkers interactive GmbH System for controlling mobile mobile units
DE102014110265A1 (en) 2014-07-22 2016-01-28 Vorwerk & Co. Interholding Gmbh Method for cleaning or processing a room by means of a self-moving device
JP6626879B2 (en) 2014-08-06 2019-12-25 アルフレッド ケルヒャー エスエー ウント コンパニー カーゲー Method for treating a floor surface and floor treatment apparatus
US10211191B2 (en) 2014-08-06 2019-02-19 Pixart Imaging Inc. Image module package with transparent sub-assembly
DE102014111217A1 (en) 2014-08-06 2016-02-11 Vorwerk & Co. Interholding Gmbh Floor cleaning device for dry and damp cleaning and method for operating a self-propelled floor cleaning device
KR102306709B1 (en) 2014-08-19 2021-09-29 삼성전자주식회사 Robot cleaner, control apparatus, control system, and control method of robot cleaner
WO2016028021A1 (en) 2014-08-20 2016-02-25 삼성전자주식회사 Cleaning robot and control method therefor
KR102527645B1 (en) 2014-08-20 2023-05-03 삼성전자주식회사 Cleaning robot and controlling method thereof
JP6325946B2 (en) 2014-08-27 2018-05-16 東芝ライフスタイル株式会社 Autonomous vehicle
JP5819498B1 (en) 2014-08-27 2015-11-24 シャープ株式会社 Autonomous mobile body and autonomous mobile body system
DE102014113040A1 (en) 2014-09-10 2016-03-10 Miele & Cie. Kg Method for operating a household appliance system
CN107000207B (en) 2014-09-24 2021-05-04 三星电子株式会社 Cleaning robot and method of controlling the same
WO2016048077A1 (en) 2014-09-24 2016-03-31 삼성전자주식회사 Cleaning robot and method for controlling cleaning robot
CN105527961A (en) 2014-09-30 2016-04-27 科沃斯机器人有限公司 Self-propelled surface-traveling robot system and method for returning to primary charging base
US9798328B2 (en) 2014-10-10 2017-10-24 Irobot Corporation Mobile robot area cleaning
CN104765362B (en) 2014-11-07 2017-09-29 深圳市银星智能科技股份有限公司 A kind of local cleaning modes of clean robot
WO2016077703A1 (en) 2014-11-13 2016-05-19 Worcester Polytechnic Institute Gyroscope assisted scalable visual simultaneous localization and mapping
US9788698B2 (en) * 2014-12-10 2017-10-17 Irobot Corporation Debris evacuation for cleaning robots
US10444756B2 (en) 2014-12-11 2019-10-15 Husqvarna Ab Navigation for a robotic working tool
US9420741B2 (en) 2014-12-15 2016-08-23 Irobot Corporation Robot lawnmower mapping
JP6532530B2 (en) 2014-12-16 2019-06-19 アクチエボラゲット エレクトロルックス How to clean a robot vacuum cleaner
EP3045936A1 (en) 2015-01-13 2016-07-20 XenomatiX BVBA Surround sensing system with telecentric optics
KR101640706B1 (en) 2015-01-28 2016-07-18 엘지전자 주식회사 Vacuum cleaner
KR102404258B1 (en) 2015-02-06 2022-06-02 삼성전자주식회사 Apparatus for returning of robot and returning method thereof
CN104634601B (en) 2015-02-09 2017-07-25 杭州市质量技术监督检测院 The detection means and method of clean robot clean-up performance
US9717387B1 (en) 2015-02-26 2017-08-01 Brain Corporation Apparatus and methods for programming and training of robotic household appliances
US9630319B2 (en) * 2015-03-18 2017-04-25 Irobot Corporation Localization and mapping using physical features
US9868211B2 (en) 2015-04-09 2018-01-16 Irobot Corporation Restricting movement of a mobile robot
DE102015006014A1 (en) 2015-05-13 2016-11-17 Universität Bielefeld Soil cultivation device and method for its navigation and swarm of tillage equipment and methods for their joint navigation
CN105045098B (en) 2015-05-29 2017-11-21 希美埃(芜湖)机器人技术有限公司 A kind of control method of Control During Paint Spraying by Robot track automatic creation system
US9919425B2 (en) * 2015-07-01 2018-03-20 Irobot Corporation Robot navigational sensor system
EP3156873B2 (en) 2015-10-15 2023-04-05 Honda Research Institute Europe GmbH Autonomous vehicle with improved simultaneous localization and mapping function
DE102015119501A1 (en) 2015-11-11 2017-05-11 RobArt GmbH Subdivision of maps for robot navigation
CN105990876B (en) 2015-12-21 2019-03-01 小米科技有限责任公司 Charging pile, identification method and device thereof and automatic cleaning equipment
CN105467398B (en) 2015-12-31 2018-08-21 上海思岚科技有限公司 Scan distance-measuring equipment
CN105527619B (en) 2016-02-05 2018-07-17 上海思岚科技有限公司 A kind of Laser Distance Measuring Equipment
DE102016102644A1 (en) 2016-02-15 2017-08-17 RobArt GmbH Method for controlling an autonomous mobile robot

Patent Citations (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE4340771A1 (en) * 1993-06-08 1994-12-15 Samsung Electronics Co Ltd Automatic cleaning device
CN1299083A (en) * 1999-10-29 2001-06-13 索尼公司 Robot and charging system, charging station searching method, connector and electric connection structure
JP2004283956A (en) * 2003-03-20 2004-10-14 Sony Corp Charging system, robot device, charging device, and charging method
US20060238159A1 (en) * 2005-04-25 2006-10-26 Lg Electronics Inc. Position calculation system for mobile robot and charging-stand return and method using the same
JP2009101044A (en) * 2007-10-25 2009-05-14 Panasonic Corp Self-propelled apparatus
KR101484941B1 (en) * 2009-06-12 2015-01-23 삼성전자주식회사 Robot cleaner and control method thereof
US20110130874A1 (en) * 2009-11-27 2011-06-02 Chuan-Ching Tsao Moving devices and controlling methods therefor
JP2015534048A (en) * 2012-08-27 2015-11-26 アクティエボラゲット エレクトロラックス Robot positioning system
WO2014043732A1 (en) * 2012-09-24 2014-03-27 RobArt GmbH Robot and method for autonomous inspection or processing of floor areas
EP2749194B1 (en) * 2012-12-26 2018-11-14 LG Electronics, Inc. Automatic cleaner
US20140278252A1 (en) * 2013-03-14 2014-09-18 Ecolab Usa Inc. System and method for monitoring of floor conditions
CN105629971A (en) * 2014-11-03 2016-06-01 贵州亿丰升华科技机器人有限公司 Robot automatic charging system and control method therefor
CN105700522A (en) * 2014-11-11 2016-06-22 沈阳新松机器人自动化股份有限公司 Robot charging method and charging system thereof
US20160147406A1 (en) * 2014-11-25 2016-05-26 Samsung Electronics Co., Ltd. Method for providing graphical user interface and electronic device for supporting the same
KR20160097966A (en) * 2015-02-10 2016-08-18 에브리봇 주식회사 A robot cleaner and a method for operating it
DE112017003497T5 (en) * 2016-07-11 2019-05-02 Groove X, Inc. Self-acting robot with controlled amount of activity

Also Published As

Publication number Publication date
JP2021180002A (en) 2021-11-18
EP3968117A2 (en) 2022-03-16
EP3968117A3 (en) 2022-06-29
US11188086B2 (en) 2021-11-30
CN114740852A (en) 2022-07-12
DE102015114883A1 (en) 2017-03-09
JP6971223B2 (en) 2021-11-24
EP3345065B1 (en) 2021-11-03
US20180246518A1 (en) 2018-08-30
WO2017037257A1 (en) 2017-03-09
CN108700880A (en) 2018-10-23
EP3345065A1 (en) 2018-07-11
JP2018526748A (en) 2018-09-13

Similar Documents

Publication Publication Date Title
US20220083070A1 (en) Identification And Localization Of A Base Station Of An Autonomous Mobile Robot
US11960304B2 (en) Localization and mapping using physical features
AU2017402126B2 (en) Vacuum cleaner and control method thereof
EP3104194B1 (en) Robot positioning system
US10423163B2 (en) Mobile robot and method of controlling same
JP6054425B2 (en) How to perform self-location estimation automatically
US20200409382A1 (en) Intelligent cleaning robot
US10213082B2 (en) Robot cleaner
CN113365535B (en) Optical marker for communicating information to autonomous device
US20160144512A1 (en) Robot cleaner and method for auto-correcting 3d sensor of the robot cleaner
KR101786516B1 (en) Moving robot and controlling method thereof
EP3738495B1 (en) Robotic vacuum cleaner and control method therefor
KR102329674B1 (en) Method of configuring position based on identification of fixed object and moving object and robot implementing thereof
EP3478143B1 (en) Robot cleaner
KR20210015126A (en) Moving Robot
KR102203438B1 (en) a Moving robot and Controlling method for the moving robot
EP3778149A1 (en) Mobile robot and mobile robot system

Legal Events

Date Code Title Description
AS Assignment

Owner name: ROBART GMBH, AUSTRIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ARTES, HAROLD;LENHARDT, FABIAN;FREUDENTHALER, CHRISTOPH;AND OTHERS;SIGNING DATES FROM 20180413 TO 20180414;REEL/FRAME:058230/0167

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION