EP1934870A2 - System und verfahren für bildabbildung und visuelle aufmerksamkeit - Google Patents

System und verfahren für bildabbildung und visuelle aufmerksamkeit

Info

Publication number
EP1934870A2
EP1934870A2 EP06816851A EP06816851A EP1934870A2 EP 1934870 A2 EP1934870 A2 EP 1934870A2 EP 06816851 A EP06816851 A EP 06816851A EP 06816851 A EP06816851 A EP 06816851A EP 1934870 A2 EP1934870 A2 EP 1934870A2
Authority
EP
European Patent Office
Prior art keywords
robot
attentional
ses
image
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
EP06816851A
Other languages
English (en)
French (fr)
Other versions
EP1934870A4 (de
Inventor
Richard Alan Ii Peters
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Vanderbilt University
Original Assignee
Vanderbilt University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Vanderbilt University filed Critical Vanderbilt University
Publication of EP1934870A2 publication Critical patent/EP1934870A2/de
Publication of EP1934870A4 publication Critical patent/EP1934870A4/de
Withdrawn legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05DSYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
    • G05D1/00Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
    • G05D1/02Control of position or course in two dimensions
    • G05D1/021Control of position or course in two dimensions specially adapted to land vehicles
    • G05D1/0231Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means
    • G05D1/0246Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means using a video camera in combination with image processing means
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05DSYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
    • G05D1/00Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
    • G05D1/02Control of position or course in two dimensions
    • G05D1/021Control of position or course in two dimensions specially adapted to land vehicles
    • G05D1/0268Control of position or course in two dimensions specially adapted to land vehicles using internal positioning means
    • G05D1/0274Control of position or course in two dimensions specially adapted to land vehicles using internal positioning means using mapping information stored in a memory device
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05BCONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
    • G05B2219/00Program-control systems
    • G05B2219/30Nc systems
    • G05B2219/35Nc in input of data, input till input file format
    • G05B2219/35144Egosphere: spherical shell 2-5-D around robot, objects are projected on it

Definitions

  • An allocentric world model places objects in a coordinate grid that does not change with the robot's position.
  • An ego-centric model is always centered on the present position of the robot.
  • An ego-centric model is described in Albus, J. S., "Outline for a theory of intelligence", IEEE Trans. Syst. Man, and Cybern., vol. 21, no. 3, 1991.
  • Albus describes an Ego-Sphere wherein the robot's environment is projected onto a spherical surface centered on the robot's current position.
  • the Ego-Sphere is a dense representation of the world in the sense that all sensory information is projected onto the Ego-Sphere.
  • Albus' Ego-Sphere is also continuous because the projection is affine.
  • the frequency is determined by the number of vertices that connect the center of one pentagon to the center of another pentagon, all pentagons being distributed on the dome evenly.
  • the SES has a tessellation of 14 and, therefore, 1963 nodes.
  • the SES facilitates the detection of events in the environment that simultaneously stimulate multiple sensors.
  • Each sensor on the robot sends information to one or more sensory processing modules (SPMs) designed to extract specific information from the data stream associated with that sensor.
  • SPMs are independent of each other and run continuously and concurrently on preferably different processors.
  • Each SPM sends information messages to an SES manager agent which stores the data, including directional sensory information if available, in the SES.
  • An object 350 is projected onto the SES by a ray 355 connecting the center 301 to the object 350.
  • Ray 355 intersects a face 360 at a point 357 defined by azimuthal angle, ⁇ s , and elevation (or polar) angle, ⁇ s .
  • Information about the object 350, such as ⁇ s and ⁇ s are stored at the vertex 370 that is closest to point 357.
  • a vertex may also contain links to behaviors stored in the DBAM.
  • Landmark mapping agents may also write to the SES, storing a pointer to an object descriptor at the vertex where the object is expected.
  • Objects may be tracked during robot movement on the SES using transformations such as those described in Peters, R. A. II, K. E. Hambuchen, K. Kawamura, and D. M. Wilkes, "The Sensory Ego-Sphere as a Short-Term Memory for Humanoids", Proc. IEEE-RAS Int'l. Conf. on Humanoid Robots, pp. 451-459, Waseda University, Tokyo, Japan, Nov. 22-24, 2001 herein incorporated by reference in its entirety.
  • Each of the behaviors linked to the current behavior computes the vector-space distance between the current state and its own pre-condition state.
  • Each behavior propagates an inhibitory signal (by adding a negative number to the activation term) that is inversely proportional to the computed distance to the other linked behaviors.
  • the propagation of the inhibitory signal between the linked behaviors has the effect that, in most instances, the behavior with the highest activation term is also the behavior whose pre-condition state most closely matches the current state of the robot.
  • the links between behaviors are created by the SAN agent during task planning but may also be created by a dream agent during the dream state. The links are task dependent and different behaviors may be linked together depending on the assigned goal.
  • the spreading activation network (SAN) agent constructs a sequence of behaviors that will take the robot from its current state to the goal state (active map) in the DBAM by back-propagating from the goal state to the current state. For each behavior added to the active map, the SAN agent performs a search for behaviors that have a pre-condition state close to the post-condition state of the added behavior and adds a link connecting the close behavior to the added behavior. An activation term characterizing the link and based on the inverse vector space distance between the linked behaviors is also added to the added behavior. The SAN agent may create several paths connecting the current state to the goal state.
  • the problem of attention arises once the SES is populated with dense information. Because of limited computational resources, only regions of interest-determined by safety, opportunity, and by the task—can be attended to, if the robot is to interact with a human-centered environment in real time. The problem lies in how to perform attention processing given a populated SES and an image input stream. There are at least two possibilities. One is to perform visual attention processing on the entire SES. The other is to detect points of interest within the individual images and combine them with the imagery that is already present.
  • the summed activation image (paragraph 88) appears better-suited for attention deployment on the SES. Processing the entire reconstructed scene image makes less information available than the summed activation image since only one image determines the most salient locations in the scene as opposed to a sequence of overlapping images. Moreover, updating the salience distribution on the SES as new information is made available is straightforward if the summed activation image is implemented. For example, this can be done simply by processing new images and combining the new attention points found with the attentional points already present. The activation at each node could be weighed by the age of each attentional point, giving more weight to newer points.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Aviation & Aerospace Engineering (AREA)
  • Remote Sensing (AREA)
  • General Physics & Mathematics (AREA)
  • Automation & Control Theory (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Electromagnetism (AREA)
  • Manipulator (AREA)
  • Image Analysis (AREA)
EP06816851A 2005-10-11 2006-10-11 System und verfahren für bildabbildung und visuelle aufmerksamkeit Withdrawn EP1934870A4 (de)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US72603305P 2005-10-11 2005-10-11
PCT/US2006/040040 WO2007044891A2 (en) 2005-10-11 2006-10-11 System and method for image mapping and visual attention

Publications (2)

Publication Number Publication Date
EP1934870A2 true EP1934870A2 (de) 2008-06-25
EP1934870A4 EP1934870A4 (de) 2010-03-24

Family

ID=37943550

Family Applications (1)

Application Number Title Priority Date Filing Date
EP06816851A Withdrawn EP1934870A4 (de) 2005-10-11 2006-10-11 System und verfahren für bildabbildung und visuelle aufmerksamkeit

Country Status (4)

Country Link
EP (1) EP1934870A4 (de)
JP (1) JP2009517225A (de)
CA (2) CA2625805C (de)
WO (1) WO2007044891A2 (de)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2002081156A2 (en) * 2001-04-06 2002-10-17 Vanderbilt University Architecture for robot intelligence
US20050149227A1 (en) * 2003-12-31 2005-07-07 Peters Richard A.Ii Architecture for robot intelligence
US20050223176A1 (en) * 2003-12-30 2005-10-06 Peters Richard A Ii Sensory ego-sphere: a mediating interface between sensors and cognition

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
FR2648071B1 (fr) * 1989-06-07 1995-05-19 Onet Procede et appareil autonomes de nettoyage automatique de sol par execution de missions programmees
US5548511A (en) * 1992-10-29 1996-08-20 White Consolidated Industries, Inc. Method for controlling self-running cleaning apparatus
US5995884A (en) * 1997-03-07 1999-11-30 Allen; Timothy P. Computer peripheral floor cleaning system and navigation method
JP4409035B2 (ja) * 2000-03-22 2010-02-03 本田技研工業株式会社 画像処理装置、特異箇所検出方法、及び特異箇所検出プログラムを記録した記録媒体
JP2002006784A (ja) * 2000-06-20 2002-01-11 Mitsubishi Electric Corp 浮遊型ロボット
JP2004086401A (ja) * 2002-08-26 2004-03-18 Sony Corp 対象物認識装置および方法

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2002081156A2 (en) * 2001-04-06 2002-10-17 Vanderbilt University Architecture for robot intelligence
US20020169733A1 (en) * 2001-04-06 2002-11-14 Peters Richard Alan Architecture for robot intelligence
US20050223176A1 (en) * 2003-12-30 2005-10-06 Peters Richard A Ii Sensory ego-sphere: a mediating interface between sensors and cognition
US20050149227A1 (en) * 2003-12-31 2005-07-07 Peters Richard A.Ii Architecture for robot intelligence

Non-Patent Citations (8)

* Cited by examiner, † Cited by third party
Title
K. A. Hambuchen: "Multi-modal attention and event binding in humanoid robot using a sensory ego-sphere" 23 February 2004 (2004-02-23), Vanderbilt university , XP007911649 * abstract * *
K. ACHIM FLEMING, R. A. PETERS, R. E. BODENHEIMER: "Image mapping and visual attention on a sensory ego-sphere" PROCEEDINGS OF THE 2006 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS'06), 9 October 2006 (2006-10-09), pages 241-246, XP031006466 *
K. ACHIM: "Chapter 1: introduction" IMAGE MAPPING AND VISUAL ATTENTION ON A SENSORY EGO-SPHERE, 18 July 2005 (2005-07-18), XP007911606 -& K. ACHIM: "Chapter 2: background material and previous work" IMAGE MAPPING AND VISUAL ATTENTION ON A SENSORY EGO-SPHERE, 18 July 2005 (2005-07-18), XP007911607 -& K. ACHIM: "Chapter 3: research system" IMAGE MAPPING AND VISUAL ATTENTION ON A SENSORY EGO-SPHERE, 18 July 2005 (2005-07-18), XP007911608 -& K. ACHIM: "Chapter 4: methods" IMAGE MAPPING AND VISUAL ATTENTION ON A SENSORY EGO-SPHERE, 18 July 2005 (2005-07-18), XP007911609 -& K. ACHIM: "Chapter 5: results" IMAGE MAPPING AND VISUAL ATTENTION ON A SENSORY EGO-SPHERE, 18 July 2005 (2005-07-18), XP007911611 -& K. ACHIM: "Chapter 6: conclusions and future work" IMAGE MAPPING AND VISUAL ATTENTION ON A SENSORY EGO-SPHERE, 18 July 2005 (2005-07-18), XP007911610 *
P. A. Fleming: "Implementing a robust 3-dimensional egocentric navigation system" 21 July 2005 (2005-07-21), Vanderbilt university , XP007911651 * abstract * *
R. A. PETERS II: "Robot behavior acquisition phase II: sequencing and compositing of behaviors learned through teleoperation" FINAL REPORT FOR NASA GRANT NNJ04HI19G, 5 October 2005 (2005-10-05), XP007911653 *
S. M. Gordon: "Cognitive map generation for localization and navigation from 3-D laser scan images" 31 March 2005 (2005-03-31), Vanderbilt university , XP007911650 * abstract * *
See also references of WO2007044891A2 *
W. Dodd: "The design of procedural, semantic, and episodic memory systems for a cognitive robot" 9 August 2005 (2005-08-09), Vanderbilt university , XP007911652 * abstract * *

Also Published As

Publication number Publication date
EP1934870A4 (de) 2010-03-24
JP2009517225A (ja) 2009-04-30
WO2007044891A2 (en) 2007-04-19
CA2625805A1 (en) 2007-04-19
CA2868135A1 (en) 2007-04-19
WO2007044891A3 (en) 2007-07-12
CA2625805C (en) 2014-11-25

Similar Documents

Publication Publication Date Title
US7835820B2 (en) System and method for image mapping and visual attention
US7328196B2 (en) Architecture for multiple interacting robot intelligences
US6697707B2 (en) Architecture for robot intelligence
JP4609584B2 (ja) ロボット装置、顔認識方法及び顔認識装置
US10131052B1 (en) Persistent predictor apparatus and methods for task switching
Shabbir et al. A survey of deep learning techniques for mobile robot applications
US20230154015A1 (en) Virtual teach and repeat mobile manipulation system
US11887363B2 (en) Training a deep neural network model to generate rich object-centric embeddings of robotic vision data
CA2625805C (en) System and method for image mapping and visual attention
Peters et al. System and method for image mapping and visual attention
Peters et al. Apparatus for multiprocessor-based control of a multiagent robot
Peters et al. Architecture for robot intelligence
AU2002258757A1 (en) Architecture for robot intelligence
WO2023100282A1 (ja) データ生成システム、モデル生成システム、推定システム、学習済みモデルの製造方法、ロボット制御システム、データ生成方法、およびデータ生成プログラム
Joshi Antipodal Robotic Grasping using Deep Learning
Bianco et al. Biologically-inspired visual landmark learning for mobile robots
Hambuchen Multi-modal attention and event binding in humanoid robot using a sensory ego-sphere
Blackburn et al. Robotic Sensor-Motor Transformations
Liu et al. Basic behavior acquisition based on multisensor integration of a robot head
Pence Autonomous Mobility and Manipulation of a 9-DoF WMRA

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 20080421

AK Designated contracting states

Kind code of ref document: A2

Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LI LT LU LV MC NL PL PT RO SE SI SK TR

A4 Supplementary search report drawn up and despatched

Effective date: 20100219

17Q First examination report despatched

Effective date: 20100609

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN

18D Application deemed to be withdrawn

Effective date: 20101221