EP3361716B1 - Robotic cleaning device - Google Patents
Robotic cleaning device Download PDFInfo
- Publication number
- EP3361716B1 EP3361716B1 EP18164153.1A EP18164153A EP3361716B1 EP 3361716 B1 EP3361716 B1 EP 3361716B1 EP 18164153 A EP18164153 A EP 18164153A EP 3361716 B1 EP3361716 B1 EP 3361716B1
- Authority
- EP
- European Patent Office
- Prior art keywords
- robotic device
- cleaning device
- robotic cleaning
- operative
- transmitting
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Lifetime
Links
- 238000004140 cleaning Methods 0.000 title claims description 80
- 230000003542 behavioural effect Effects 0.000 claims description 74
- 238000001514 detection method Methods 0.000 claims description 62
- 230000033001 locomotion Effects 0.000 claims description 41
- 230000000694 effects Effects 0.000 claims description 18
- 230000000712 assembly Effects 0.000 claims description 13
- 238000000429 assembly Methods 0.000 claims description 13
- 238000010408 sweeping Methods 0.000 claims description 9
- 230000005855 radiation Effects 0.000 claims description 6
- 230000005540 biological transmission Effects 0.000 claims description 5
- 230000008859 change Effects 0.000 claims description 4
- 238000000034 method Methods 0.000 description 24
- 239000013598 vector Substances 0.000 description 15
- 230000007704 transition Effects 0.000 description 14
- 230000004044 response Effects 0.000 description 13
- 230000008569 process Effects 0.000 description 12
- 238000012545 processing Methods 0.000 description 12
- 230000000875 corresponding effect Effects 0.000 description 10
- 230000006870 function Effects 0.000 description 9
- 230000004913 activation Effects 0.000 description 8
- CIWBSHSKHKDKBQ-JLAZNSOCSA-N Ascorbic acid Chemical compound OC[C@H](O)[C@H]1OC(=O)C(O)=C1O CIWBSHSKHKDKBQ-JLAZNSOCSA-N 0.000 description 6
- 230000006399 behavior Effects 0.000 description 4
- 230000001360 synchronised effect Effects 0.000 description 4
- 230000002596 correlated effect Effects 0.000 description 3
- 238000004364 calculation method Methods 0.000 description 2
- 230000000295 complement effect Effects 0.000 description 2
- 230000003247 decreasing effect Effects 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 230000005670 electromagnetic radiation Effects 0.000 description 2
- 230000001771 impaired effect Effects 0.000 description 2
- 230000004807 localization Effects 0.000 description 2
- 230000007246 mechanism Effects 0.000 description 2
- 238000001228 spectrum Methods 0.000 description 2
- 230000009471 action Effects 0.000 description 1
- 230000002411 adverse Effects 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 230000004888 barrier function Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 238000007796 conventional method Methods 0.000 description 1
- 230000001351 cycling effect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000006073 displacement reaction Methods 0.000 description 1
- 239000000428 dust Substances 0.000 description 1
- 230000007274 generation of a signal involved in cell-cell signaling Effects 0.000 description 1
- 230000000977 initiatory effect Effects 0.000 description 1
- 238000012423 maintenance Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000012913 prioritisation Methods 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 238000000926 separation method Methods 0.000 description 1
- 230000028838 turning behavior Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0268—Control of position or course in two dimensions specially adapted to land vehicles using internal positioning means
- G05D1/0272—Control of position or course in two dimensions specially adapted to land vehicles using internal positioning means comprising means for registering the travel distance, e.g. revolutions of wheels
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S5/00—Position-fixing by co-ordinating two or more direction or position line determinations; Position-fixing by co-ordinating two or more distance determinations
- G01S5/02—Position-fixing by co-ordinating two or more direction or position line determinations; Position-fixing by co-ordinating two or more distance determinations using radio waves
- G01S5/12—Position-fixing by co-ordinating two or more direction or position line determinations; Position-fixing by co-ordinating two or more distance determinations using radio waves by co-ordinating position lines of different shape, e.g. hyperbolic, circular, elliptical or radial
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S5/00—Position-fixing by co-ordinating two or more direction or position line determinations; Position-fixing by co-ordinating two or more distance determinations
- G01S5/16—Position-fixing by co-ordinating two or more direction or position line determinations; Position-fixing by co-ordinating two or more distance determinations using electromagnetic waves other than radio waves
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0212—Control of position or course in two dimensions specially adapted to land vehicles with means for defining a desired trajectory
- G05D1/0225—Control of position or course in two dimensions specially adapted to land vehicles with means for defining a desired trajectory involving docking at a fixed facility, e.g. base station or loading bay
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0231—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means
- G05D1/0242—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means using non-visible light signals, e.g. IR or UV signals
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0268—Control of position or course in two dimensions specially adapted to land vehicles using internal positioning means
- G05D1/0274—Control of position or course in two dimensions specially adapted to land vehicles using internal positioning means using mapping information stored in a memory device
Definitions
- the present invention relates generally to mobile robotic devices, and more particularly, to a navigational control system for a robotic device operating under the direction of a navigation control algorithm that tracks the movement activity of the robotic device and utilizes movement activity to provide one or more control instructions to the robotic device, which in response thereto, implements a prescribed conduct, e.g., one or more basic maneuvers and/or behavioral patterns to increase the coverage efficiency of such robotic devices for covering a defined working area, e.g., sweeping, vacuuming a room, and/or performing residential yard maintenance.
- a prescribed conduct e.g., one or more basic maneuvers and/or behavioral patterns to increase the coverage efficiency of such robotic devices for covering a defined working area, e.g., sweeping, vacuuming a room, and/or performing residential yard maintenance.
- Robotic engineers have long worked on developing an effective method of autonomous cleaning. This has led to the development of two separate and distinct schemes for autonomous robotic devices: (1) deterministic cleaning; and (2) random cleaning.
- the autonomous robotic device follows a defined path, e.g., a boustrophedon path that is calculated to facilitate complete cleaning coverage of a given area while eliminating redundant cleaning.
- Deterministic cleaning requires that the robotic device maintain precise position knowledge at all times, as well as its position history (where it has been), which, in turn, requires a sophisticated positioning system.
- a suitable positioning system - a positioning system suitably accurate for deterministic cleaning might rely on scanning laser ranging systems, ultrasonic transducers, a carrier phase differential GPS, or other sophisticated methods - is typically prohibitively expensive and labor intensive, requiring an involved pre-setup to accommodate the unique conditions of each area to be cleaned, e.g., room geometry, furniture locations.
- methods that rely on global positioning are typically incapacitated by failure of any part of the positioning system.
- the RoboScrub device built by Denning Mobile Robotics and Windsor Industries.
- the RoboScrub device employs sonar and infrared detectors, bump sensors, and a high-precision laser navigation system to define the deterministic cleaning path.
- the navigation system employed with the RoboScrub device requires numerous large bar code targets to be set up in various strategic positions within the area to be cleaned, and effective operation of the navigation system requires that at least four of such targets be visible simultaneously. This target accessibility requirement effectively limits the use of the RoboScrub device to large uncluttered open areas.
- Dead reckoning consists of continually measuring the precise rotation of each drive wheel (e.g., using optical shaft encoders) to continually calculate the current position of the robotic device, based upon a known starting point and orientation.
- the drive wheels of dead reckoning robotic devices are almost always subject to some degree of slippage, which leads to errors in the calculation of current position. Accordingly, dead reckoning robotic devices are generally considered unreliable for cleaning operations of any great duration - resulting in intractable system neglect, i.e., areas of the surface to be cleaned are not cleaned.
- Other representative examples of pseudo-deterministic robotic devices are described in U.S. Patent Nos. 6,255,793 (Peless et al. ) and 5,109,566 (Kobayashi et al. ).
- a robotic device operating in random motion, under the control of one or more random-motion algorithms stored in the robotic device, represents the other basic approach to cleaning operations using autonomous robotic devices.
- the robotic device autonomously implement such random-motion algorithm(s) in response to internal events, e.g., signals generated by a sensor system, elapse of a time period (random or predetermined).
- internal events e.g., signals generated by a sensor system
- a robotic device operating under the control of a random-motion algorithm will provide acceptable cleaning coverage given enough cleaning time.
- a robotic device utilizing a random-motion algorithm must operate for a longer period of time to achieve acceptable cleaning coverage.
- the random-motion robotic device must run approximately five times longer than a deterministic robotic device having similarly sized cleaning mechanisms and moving at approximately the same speed.
- an area to be cleaned that includes one or more randomly-situated obstacles causes a marked increase in the running time for a random-motion robotic device to effect 98% cleaning coverage. Therefore, while a random motion robotic device is a relatively inexpensive means of cleaning a defined working area as contrasted to a deterministic robotic device, the random-motion robotic device requires a significantly higher cleaning time.
- One aim is to provide a navigational control system that enhances the cleaning efficiency of a robotic device by adding a deterministic component (in the form of a conduct prescribed by a navigation control algorithm) to the random motion of the robotic device generated by predetermined behavioral modes stored in the robotic device.
- a deterministic component in the form of a conduct prescribed by a navigation control algorithm
- Another aim is to provide a navigational control unit operating under a navigation control algorithm that includes a predetermined triggering event that defines when the prescribed conduct will be implemented by the robotic device.
- a navigational control system for deterministically altering movement activity of a robotic device operating in a defined working area, comprising a transmitting subsystem integrated in combination with the robotic device, the transmitting subsystem comprising means for emitting a number of directed beams, each directed beam having a predetermined emission pattern, and a receiving subsystem functioning as a base station that includes a navigation control algorithm that defines a predetermined triggering event for the navigational control system and a set of detection units positioned within the defined working area, the detection units being positioned in a known aspectual relationship with respect to one another, the set of detection units being configured and operative to detect one or more of the directed beams emitted by the transmitting system; and wherein the receiving subsystem is configured and operative to process the one or more detected directed beams under the control of the navigational control algorithm to determine whether the predetermined triggering event has occurred, and, if the predetermined triggering event has occurred transmit a control signal to the robotic device, wherein reception of the control signal by the robotic device causes the
- Figure 1 is a top-view schematic of an exemplary preferred embodiment of a robotic device 100 having particular utility in combination with a navigational control system 10.
- Figure 2 is a block diagram of the hardware of the robot device 100 of Figure 1 .
- the hardware and behavioral modes (coverage behaviors for cleaning operations; escape behaviors for transitory movement patterns; and safety behaviors for emergency conditions) of the robotic device 100 which is manufactured, distributed, and/or marketed by the iRobot Corporation of Burlington, MA under the ROOMBA trademark, are briefly described in the following paragraphs to facilitate a more complete understanding of the navigational control system 10 of the present invention. Further details regarding the hardware and behavioral modes of the robotic device 100 can be found in commonly-owned, co-pending U.S. nonprovisional patent application Serial No. 10/167,851, filed 12 June 2002 , entitled METHOD AND SYSTEM FOR MULTI-MODE COVERAGE FOR AN AUTONOMOUS ROBOT, and U.S. nonprovisional patent application Serial No. 10/320,729, filed 16 December 2002 , entitled AUTONOMOUS FLOOR-CLEANING DEVICE.
- forward refers to the primary direction of motion (forward) of the robotic device (see arrow identified by reference character “FM” in Figure 1 ).
- the fore/aft axis FA X of the robotic device 100 coincides with the medial diameter of the robotic device 100 that divides the robotic device 100 into generally symmetrical right and left halves, which are defined as the dominant and non-dominant sides, respectively.
- the robotic device 100 has a generally cylindrical housing infrastructure that includes a chassis 102 and an outer shell 104 secured to the chassis 102 that define a structural envelope of minimal height (to facilitate movement under furniture).
- the hardware comprising the robotic device 100 can be generally categorized as the functional elements of a power system, a motive power system, a sensor system, a control module, a side brush assembly, or a self-adjusting cleaning head system, respectively, all of which are integrated in combination with the housing infrastructure.
- the robotic device 100 further includes a forward bumper 106 having a generally arcuate configuration and a nose-wheel assembly 108.
- the forward bumper 106 (illustrated as a single component; alternatively, a two-segment component) is integrated in movable combination with the chassis 102 (by means of displaceable support members pairs) to extend outwardly therefrom. Whenever the robotic device 100 impacts an obstacle (e.g., wall, furniture) during movement thereof, the bumper 106 is displaced (compressed) towards the chassis 102 and returns to its extended (operating) position when contact with the obstacle is terminated.
- an obstacle e.g., wall, furniture
- the nose-wheel assembly 108 is mounted in biased combination with the chassis 102 so that the nose-wheel subassembly 108 is in a retracted position (due to the weight of the robotic device 100) during cleaning operations wherein it rotates freely over the surface being cleaned.
- the nose-wheel subassembly 108 encounters a drop-off during operation (e.g., descending stairs, split-level floors)
- the nose-wheel assembly 108 is biased to an extended position.
- the hardware of the power system which provides the energy to power the electrically-operated hardware of the robotic device 100, comprises a rechargeable battery pack 110 (and associated conduction lines, not shown) that is integrated in combination with the chassis 102.
- the motive power system provides the means that propels the robotic device 100 and operates the cleaning mechanisms, e.g., side brush assembly and the self-adjusting cleaning head system, during movement of the robotic device 100.
- the motive power system comprises left and right main drive wheel assemblies 112L, 112R, their associated independent electric motors 114L, 114R, and electric motors 116, 118 for operation of the side brush assembly and the self-adjusting cleaning head subsystem, respectively.
- the main drive wheel assemblies 112L, 112R are independently mounted in biased combination with the chassis 102 (for pivotal motion with respect thereto) at opposed ends of the transverse diameter (with respect to the fore-aft axis FA X ) of the robotic device 100 and are in a retracted position (due to the weight of the robotic device 100) during operation thereof wherein the axes of rotation are approximately coplanar with the bottom of the chassis 102. If the robotic device 100 is removed from the surface being cleaned, the main wheel assemblies 112L, 112R are pivotally-biased to an extended position wherein their axes of rotation are below the bottom plane of the chassis 102 (in this extended position the rechargeable battery pack 110 is automatically turned off by the control module executing one of the safety behavioral modes).
- the electric motors 114L, 114R are mechanically coupled to the main drive wheel assemblies 112L, 112R, respectively, and independently operated by control signals generated by the control module as a response to the implementation of a behavioral mode. Independent operation of the electric motors 114L, 114R allows the main wheel assemblies 112L, 112R to be: (1) rotated at the same speed in the same direction to propel the robotic device 100 in a straight line, forward or aft; (2) differentially rotated (including the condition wherein one wheel assembly is not rotated) to effect a variety of right and/or left turning patterns (over a spectrum of sharp to shallow turns) for the robotic device 100; and (3) rotated at the same speed in opposite directions to cause the robotic device 100 to turn in place, i.e., "spin on a dime", to provide an extensive repertoire of movement capability for the robotic device 100.
- the sensor system comprises a variety of different sensor units that are operative to generate signals that control the behavioral mode operations of the robotic device 100.
- the described robotic device 100 includes obstacle detection units 120, cliff detection units 122, wheel drop sensors 124, an obstacle-following unit 126, a virtual wall omnidirectional detector 128, stall-sensor units 130, and main wheel encoder units 132.
- the obstacle (“bump”) detection units 120 are IR break beam sensors mounted in combination with the displaceable support member pairs of the forward bumper 106. These detection units 120 are operative to generate one or more signals indicating relative displacement between one or more support member pairs whenever the robotic device 100 impacts an obstacle such that the forward bumper 106 is compressed. These signals are processed by the control module to determine an approximate point of contact with the obstacle relative to the fore-aft axis FA X of the robotic device 100 (and the behavioral mode(s) to be implemented).
- the cliff detection units 122 are mounted in combination with the forward bumper 106.
- Each cliff detection unit 122 comprises an IR emitter - detector pair configured and operative to establish a focal point such that radiation emitted downwardly by the emitter is reflected from the surface being traversed and detected by the detector. If reflected radiation is not detected by the detector, i.e., a drop-off is encountered, the cliff detection unit 122 transmits a signal to the control module (which causes one or more behavioral modes to be implemented).
- a wheel drop sensor 124 such as a contact switch is integrated in combination with each of the main drive wheel assemblies 112L, 112R and the nose wheel assembly 108 and is operative to generate a signal whenever any of the wheel assemblies is in an extended position, i.e., not in contact with the surface being traversed, (which causes the control module to implement one ore more behavioral modes).
- the obstacle-following unit 126 for the described embodiment is an IR emitter - detector pair mounted on the 'dominant' side (right hand side of Figure 1 ) of the robotic device 100.
- the emitter - detector pair is similar in configuration to the cliff detection units 112, but is positioned so that the emitter emits radiation laterally from the dominant side of the robotic device 100.
- the unit 126 is operative to transmit a signal to the control module whenever an obstacle is detected as a result of radiation reflected from the obstacle and detected by the detector.
- the control module in response to this signal, causes one or more behavioral modes to be implemented.
- a virtual wall detection system for use in conjunction with the described embodiment of the robotic device 100 comprises an omnidirectional detector 128 mounted atop the outer shell 104 and a stand-alone transmitting unit (not shown) that transmits an axially-directed confinement beam.
- the stand-alone transmitting unit is positioned so that the emitted confinement beam blocks an accessway to a defined working area, thereby restricting the robotic device 100 to operations within the defined working area (e.g., in a doorway to confine the robotic device 100 within a specific room to be cleaned).
- the omnidirectional detector 128 Upon detection of the confinement beam, transmits a signal to the control module (which causes one or more behavioral modes to be implemented to move the robotic device 100 away from the confinement beam generated by the stand-alone transmitting unit).
- a stall sensor unit 130 is integrated in combination with each electric motor 114L, 114R, 116, 118 and operative to transmit a signal to the control module when a change in current is detected in the associated electric motor (which is indicative of a dysfunctional condition in the corresponding driven hardware).
- the control module is operative in response to such a signal to implement one or more behavioral modes.
- An IR encoder unit 132 (see Figure 2 ) is integrated in combination with each main wheel assembly 112L, 112R and operative to detect the rotation of the corresponding wheel and transmit signals corresponding thereto the control module (wheel rotation can be used to provide an estimate of distance traveled for the robotic device 100).
- the control module comprises the microprocessing unit 135 illustrated in Figure 2 that includes I/O ports connected to the sensors and controllable hardware of the robotic device 100, a microcontroller, and ROM and RAM memory.
- the I/O ports function as the interface between the microcontroller and the sensor units and controllable hardware, transferring signals generated by the sensor units to the microcontroller and transferring control (instruction) signals generated by the microcontroller to the controllable hardware to implement a specific behavioral mode.
- the microcontroller is operative to execute instruction sets for processing sensor signals, implementing specific behavioral modes based upon such processed signals, and generating control (instruction) signals for the controllable hardware based upon implemented behavioral modes for the robotic device 100.
- the cleaning coverage and control programs for the robotic device 100 are stored in the ROM of the microprocessing unit 135, which includes the behavioral modes, sensor processing algorithms, control signal generation algorithms and a prioritization algorithm for determining which behavioral mode or modes are to be given control of the robotic device 100.
- the RAM of the microprocessing unit 135 is used to store the active state of the robotic device 100, including the ID of the behavioral mode(s) under which the robotic device 100 is currently being operated and the hardware commands associated therewith.
- the side brush assembly 140 is configured and operative to entrain particulates outside the periphery of the housing infrastructure and to direct such particulates towards the self-adjusting cleaning head system.
- the side brush assembly 140 provides the robotic device 100 with the capability of cleaning surfaces adjacent to base-boards when the robotic device is operated in an Obstacle-Following behavioral mode.
- the side brush assembly 140 is preferably mounted in combination with the chassis 102 in the forward quadrant on the dominant side of the robotic device 100.
- the self-adjusting cleaning head system 145 for the described robotic device 100 comprises a dual-stage brush assembly and a vacuum assembly, each of which is independently powered by an electric motor (reference numeral 118 in Figure 1 actually identifies two independent electric motors -- one for the brush assembly and one for the vacuum assembly).
- the cleaning capability of the robotic device 100 is commonly characterized in terms of the width of the cleaning head system 145 (see reference character W in Figure 1 ).
- the dual-stage brush assembly and the inlet of the vacuum assembly are integrated in combination with a deck structure, which is pivotally mounted in combination with the chassis 102 and operatively integrated with the motor of the dual-stage brush assembly.
- the brush assembly motor provides the motive force to pivot the deck structure with respect to the chassis 102.
- the pivoting deck structure provides the self adjusting capability for the cleaning head assembly 145, which allows the robotic device 100 to readily transition between disparate surfaces during cleaning operations, e.g., carpeted surface to bare surface or vice versa, without hanging up.
- the dual-stage brush assembly comprises asymmetric, counter-rotating brushes that are positioned (forward of the inlet of the vacuum assembly), configured and operative to direct particulate debris into a removable dust cartridge (not shown).
- the positioning, configuration, and operation of the brush assembly concomitantly directs particulate debris towards the inlet of the vacuum assembly such that particulates that are not swept up by the dual-stage brush assembly can be subsequently ingested by the vacuum assembly as a result of movement of the robotic device 100.
- Operation of the vacuum assembly independently of the self-adjustable brush assembly allows the vacuum assembly to generate and maintain a higher vacuum force using a battery-power source than would be possible if the vacuum assembly were operated in dependence with the brush assembly.
- the robotic device 100 uses a variety of behavioral modes to effectively clean a defined working area where behavioral modes are layers of control systems that can be operated in parallel.
- the microprocessor unit 135 is operative to execute a prioritized arbitration scheme to identify and implement one or more dominant behavioral modes for any given scenario based upon inputs from the sensor system.
- the behavioral modes for the described robotic device 100 can be characterized as: (1) coverage behavioral modes; (2) escape behavioral modes; and (3) safety behavioral modes.
- Coverage behavioral modes are primarily designed to allow the robotic device 100 to perform its cleaning operations in an efficient and effective manner and the escape and safety behavioral modes are priority behavioral modes implemented when a signal from the sensor system indicates that normal operation of the robotic device 100 is impaired, e.g., obstacle encountered, or is likely to be impaired, e.g., drop-off detected.
- Representative and illustrative coverage behavioral (cleaning) modes for the robotic device 100 include: (1) a Spot Coverage pattern; (2) an Obstacle-Following (or Edge-Cleaning ) Coverage pattern, and (3) a Room Coverage pattern.
- the Spot Coverage pattern causes the robotic device 100 to clean a limited area within the defined working area, e.g., a high-traffic area.
- the Spot Coverage pattern is implemented by means of a spiral algorithm (but other types of self-bounded area algorithms, e.g., polygonal, can be used).
- the spiral algorithm which causes outward spiraling (preferred) or inward spiraling movement of the robotic device 100, is implemented by control signals from the microprocessing unit 135 to the main wheel assemblies 112L, 112R to change the turn radius/radii thereof as a function of time (thereby increasing/decreasing the spiral movement pattern of the robotic device 100).
- the robotic device 100 is operated in the Spot Coverage pattern for a predetermined or random period of time, for a predetermined or random distance (e.g., a maximum spiral distance) and/or until the occurrence of a specified event, e.g., activation of one or more of the obstacle detection units 120 (collectively a transition condition).
- a transition condition occurs, the robotic device 100 can implement or transition to a different behavioral mode, e.g., a Straight Line behavioral mode (in a preferred embodiment of the robotic device 100, the Straight Line behavioral mode is a low priority, default behavior that propels the robot in an approximately straight line at a preset velocity of approximately 0.306 m/s) or a Bounce behavioral mode in combination with a Straight Line behavioral mode.
- a Straight Line behavioral mode in a preferred embodiment of the robotic device 100, the Straight Line behavioral mode is a low priority, default behavior that propels the robot in an approximately straight line at a preset velocity of approximately 0.306 m/s
- Bounce behavioral mode in combination with a
- the robotic device 100 can take other actions in lieu of transitioning to a different behavioral mode.
- the robotic device 100 can momentarily implement a behavioral mode to avoid or escape the obstacle and resume operation under control of the spiral algorithm (i.e., continue spiraling in the same direction).
- the robotic device 100 can momentarily implement a behavioral mode to avoid or escape the obstacle and resume operation under control of the spiral algorithm (but in the opposite direction - reflective spiraling).
- the Obstacle-Following Coverage pattern causes the robotic device 100 to clean the perimeter of the defined working area, e.g., a room bounded by walls, and/or the perimeter of an obstacle (e.g., furniture) within the defined working area.
- the robotic device 100 utilizes obstacle-following unit 126 to continuously maintain its position with respect to an obstacle, e.g., wall, furniture, so that the motion of the robotic device 100 causes it to travel adjacent to and concomitantly clean along the perimeter of the obstacle.
- Different embodiments of the obstacle-following unit 126 can be used to implement the Obstacle-Following behavioral pattern.
- the obstacle-following unit 126 is operated to detect the presence or absence of the obstacle. In an alternative embodiment, the obstacle-following unit 126 is operated to detect an obstacle and then maintain a predetermined distance between the obstacle and the robotic device 100.
- the microprocessing unit 135 is operative, in response to signals from the obstacle-following unit, to implement small CW or CCW turns to maintain its position with respect to the obstacle.
- the robotic device 100 implements a small CW when the robotic device 100 transitions from obstacle detection to non-detection (reflection to non-reflection) or to implement a small CCW turn when the robotic device 100 transitions from non-detection to detection (non-reflection to reflection). Similar turning behaviors are implemented by the robotic device 100 to maintain the predetermined distance from the obstacle.
- the robotic device 100 is operated in the Obstacle-Following behavioral mode for a predetermined or random period of time, for a predetermined or random distance (e.g., a maximum or minimum distance) and/or until the occurrence of a specified event, e.g., activation of one or more of the obstacle detection units 120 a predetermined number of times (collectively a transition condition).
- the microprocessor 135 will cause the robotic device to implement an Align behavioral mode upon activation of the obstacle-detection units 120 in the Obstacle-Following behavioral mode wherein the implements a minimum angle CCW turn to align the robotic device 100 with the obstacle.
- the Room Coverage pattern can be used by the robotic device 100 to clean any defined working area that is bounded by walls, stairs, obstacles or other barriers (e.g., a virtual wall unit).
- a preferred embodiment for the Room Coverage pattern comprises the Random-Bounce behavioral mode in combination with the Straight Line behavioral mode. Initially, the robotic device 100 travels under control of the Straight-Line behavioral mode, i.e., straight-line algorithm (main drive wheel assemblies 112L, 112R operating at the same rotational speed in the same direction) until an obstacle is encountered.
- the microprocessing unit 135 operative to compute an acceptable range of new directions based upon the obstacle detection unit(s) 126 activated.
- the microprocessing unit 135 selects a new heading from within the acceptable range and implements a CW or CCW turn to achieve the new heading with minimal movement.
- the new turn heading may be followed by forward movement to increase the cleaning efficiency of the robotic device 100.
- the new heading may be randomly selected across the acceptable range of headings, or based upon some statistical selection scheme, e.g., Gaussian distribution.
- the microprocessing unit 135 can be programmed to change headings randomly or at predetermined times, without input from the sensor system.
- the robotic device 100 is operated in the Room Coverage behavioral mode for a predetermined or random period of time, for a predetermined or random distance (e.g., a maximum or minimum distance) and/or until the occurrence of a specified event, e.g., activation of the obstacle-detection units 120 a predetermined number of times (collectively a transition condition).
- a predetermined or random distance e.g., a maximum or minimum distance
- a specified event e.g., activation of the obstacle-detection units 120 a predetermined number of times (collectively a transition condition).
- a preferred embodiment of the robotic device 100 includes four escape behavioral modes: a Turn behavioral mode, an Edge behavioral mode, a Wheel Dro p behavioral mode, and a Slow behavioral mode.
- a Turn behavioral mode a Turn behavioral mode
- an Edge behavioral mode a Wheel Dro p behavioral mode
- a Slow behavioral mode a preferred embodiment of the robotic device 100 includes four escape behavioral modes: a Turn behavioral mode, an Edge behavioral mode, a Wheel Dro p behavioral mode, and a Slow behavioral mode.
- Other behavioral modes can be utilized by the robotic device 100.
- One or more of these behavioral modes may be implemented, for example, in response to a current rise in one of the electric motors 116, 118 of the side brush assembly 140 or dual-stage brush assembly above a low or high stall threshold, forward bumper 106 in compressed position for determined time period, detection of a wheel-drop event.
- the robotic device 100 turns in place in a random direction, starting at higher velocity (e.g., twice normal turning velocity) and decreasing to a lower velocity (one-half normal turning velocity), i.e., small panic turns and large panic turns, respectively.
- Low panic turns are preferably in the range of 45° to 90°
- large panic turns are preferably in the range of 90° to 270°.
- the Turn behavioral mode prevents the robotic device 100 from becoming stuck on room impediments, e.g., high spot in carpet, ramped lamp base, from becoming stuck under room impediments, e.g., under a sofa, or from becoming trapped in a confined area.
- the Edge behavioral mode follows the edge of an obstacle unit it has turned through a predetermined number of degrees, e.g., 60°, without activation of any of the obstacle detection units 120, or until the robotic device has turned through a predetermined number of degrees, e.g., 170°, since initiation of the Edge behavioral mode.
- the Edge behavioral mode allows the robotic device 100 to move through the smallest possible openings to escape from confined areas.
- the microprocessor 135 In the Wheel Drop behavioral mode, the microprocessor 135 reverses the direction of the main wheel drive assemblies 112L, 112R momentarily, then stops them. If the activated wheel drop sensor 124 deactivates within a predetermined time, the microprocessor 135 then reimplements the behavioral mode that was being executed prior to the activation of the wheel drop sensor 124.
- the Slow behavioral mode is implemented to slowed down the robotic device 100 for a predetermined distance and then ramped back up to its normal operating speed.
- a safety condition e.g., a series of brush or wheel stalls that cause the corresponding electric motors to be temporarily cycled off, wheel drop sensor 124 or a cliff detection sensor 122 activated for greater that a predetermined period of time, the robotic device 100 is generally cycled to an off state.
- an audible alarm may be generated.
- Figure 3 is a schematic representation of a navigational control system 10 for use in combination with a robotic device 100 to enhance the cleaning efficiency thereof by adding a deterministic component (in the form of a control signal that remotely controls the movement of the robotic device 100) to the motion algorithms, including random motion, autonomously implemented by the robotic device 100.
- the navigational control system 10 comprises a transmitting subsystem 12 and a receiving subsystem 20 operating under the direction of a navigation control algorithm.
- the navigation control algorithm includes a definition of a predetermined triggering event.
- the specific features and characteristics of the transmitting subsystem 12 and the receiving subsystem 20 depend upon whether the particular subsystem is integrated in combination with the robotic device 100 or functions as a "base station" for the navigational control system 10.
- the navigational control system 10 is operative, under the direction of the navigation control algorithm, to monitor the movement activity of the robotic device 100 within the defined working area.
- the monitored movement activity is defined in terms of the "position history" of the robotic device 100 as described in further detail below.
- the monitored movement activity is further defined in terms of the "instantaneous position" of the robotic device 100 as defined in further detail below.
- the predetermined triggering event is a specific occurrence or condition in the movement activity of the robotic device 100.
- the navigational control system 10 is operative to generate and communicate a control signal to the robotic device 100.
- the robotic device 100 is operative to implement or execute a conduct prescribed by the control signal, i.e., the prescribed conduct. This prescribed conduct represents a deterministic component of the movement activity of the robotic device 100.
- the system 10 is configured and operative to create a "tessellation" of any defined working area where the robotic device 100 is to be operated, e.g., a room to be cleaned.
- Tessellate is used herein in the sense that the defined working area is segmented into a set of individual cells, which may or may not be of equal size.
- Figure 4 exemplarily illustrates the polar tessellation of a defined working area into a set of individual cells C (reference characters BS T identify the "base station”) of unequal size.
- each cell C (in terms of its center) is identified in terms of polar coordinates (r, ⁇ ) referenced to the base station BS T as the origin (0, 0).
- a grid map of the cells C comprising the defined working area is stored in memory of the navigation control system 10.
- coordinate systems e.g., a planar Cartesian coordinate system, can be used by the navigational control system 10 to define the position of individual cells C within the predetermined working area.
- the navigational control system 10 is operative to define the size the individual cells C so that the length and width dimensions of an individual cell C are no larger than one-half the width (W) of the cleaning head system 145 of the robotic device 100 (see Figure 1 and corresponding discussion above).
- the navigational control system 10 is operative to generate a position history of the robotic device 100 within the defined working area in terms of such individual cells C (to minimize the memory requirements for storage of position history).
- the position history comprises a set of discrete, instantaneous positions (in terms of individual cells C) of the robotic device 100 over a time interval where the time interval is a variable that depends upon the "triggering condition" of the navigation control algorithm implemented by the navigational control system 10.
- Each discrete instantaneous position of the robotic device 100 is determined by operating the transmitting subsystem 12 to emit a set of directional beams and operating the receiving subsystem 20 to detect one or more of such directional beams and process a signal parameter of the detected beam(s) to determine an absolute bearing parameter and a distance parameter between the transmitting subsystem 12 and the receiving subsystem 20 at a point in time.
- Each pair of bearing, distance parameters establishes a discrete instantaneous position for the robotic device 100.
- the navigational control system 10 is operative to correlate each discrete instantaneous position to one individual cell C of the grid map.
- a set of bearing and position pairs, i.e., a set of instantaneous positions, over a time interval defines a set of cells C, which are identified in the receiving subsystem 20 as the position history of the robotic device 100 for the time interval.
- the system 10 processes each discrete instantaneous position as it is established, under the control of the navigation control algorithm, to determine whether such discrete instantaneous position is the predetermined triggering event defined by the navigation control algorithm.
- the system 10 is additionally configured and operative to determine a travel vector (indicating the direction of motion of the robotic device 100 within an individual cell C or at the discrete instantaneous position) at each point in time. These travel vectors may be stored in memory in conjunction with the corresponding cells C as a component of the position history of the robotic device 100.
- the navigational control system 10 is further operative, under direction of the navigational control algorithm, to generate and communicate a control signal to the robotic device 100 whenever the navigational control system 100 realizes the predetermined triggering event.
- the robotic device 100 is configured and operative to initiate a prescribed conduct.
- the prescribed conduct comprises the deterministic component added to the random motion movement activity of the robotic device 100 by means of the navigational control system 10 according to the present invention.
- the prescribed conduct of the robotic device 100 comprises one or more basic maneuvers such as CW and CCW turns, forward or aft (straight line) movement, slow down, speed up, and stop.
- the CW and/or CCW turns can be implemented using the turning techniques of the robotic device 100 described above, and the turn angles can be, for example, over a 360° spectrum at predetermined intervals, e.g., 5° or 10°.
- the CW and/or CCW turns can be to a specified azimuthal heading (referenced to the base station as the origin) where the navigational control system 10 is configured and operative so that the travel vector is a determinable variable.
- forward (straight line) movement is typically the default maneuver that the robotic device 100 automatically reverts to (implements) once one or more of the other basic maneuvers has been completed.
- the prescribed conduct of the robotic device 100 comprises one or more of the behavioral modes described herein. In yet a further preferred embodiment, the prescribed conduct of the robotic device 100 comprises a combination of the basic maneuvers and the behavioral modes described herein.
- the transmitting subsystem 12 is operative to transmit a number of directed beams having a predetermined emission pattern along a specific propagation axis.
- the directed beams are planar, i.e., substantially parallel to the surface of the defined working area.
- the transmitting subsystem 12 is integrated in combination with the robotic device 100.
- the transmitting subsystem 12 is configured and operative to functionally emulate an omnidirectional transmission source with respect to the defined working area, i.e., by emitting a plurality of directed beams that cover the defined working area.
- the robotic device 100 further includes a receiver unit 16 (see Figure 3 ) configured and operative to receive control signals from the receiving subsystem 20 (see discussion below regarding the transmitting unit 32 of the receiving subsystem 20). While the receiver unit 16 is depicted as a dedicated receiving unit for the control signals, it is preferable that the omnidirectional detector 128 (of the virtual wall detection system) described above be adapted to detect and process such control signals.
- the transmitting subsystem 12 comprises a conventional mechanical sweeping transmitter, e.g., a laser, that is integrated in combination with a high point of the housing infrastructure of the robotic device 100 so that none of the structural features of the robotic device 100 interfere with the operation thereof.
- the mechanical sweeping transmitter is configured and operative to emit the plurality of directed beams while concomitantly redirecting (mechanically sweeping) the transmitting element so that each directed beam has a different propagation axis.
- Other features and characteristics of the mechanical sweeping transmitter are described below in terms of individual transmitting units 14 N for ease of description.
- the transmitting subsystem 12 comprises a set of transmitting units 14 N , where N is an integer defining the number of individual transmitting units comprising the set for the navigational control system 10, that are integrated in combination with the robotic device 100 about the periphery of its housing infrastructure.
- Each transmitting unit 14N is configured and operative to emit a directed beam having a predetermined emission pattern along a specific propagation axis.
- the transmitting subsystem 12 is configured and operative so that the emitted directed beams are planar.
- the transmitting units 14 N are fungible/interchangeable, each operating to emit a directed beam at a common operating frequency.
- the common operating frequency for the transmitting units 14 N lies in the infrared range, i.e., about 750 nm to about 1.4 x 10 4 nm, preferably about 880 nm to about 980 nm, although one skilled in the art will appreciate that other wavelengths, e.g., in the radio frequency range, microwave frequency range, can be used in the practice of the navigational control system 10.
- the common operating frequency directed beams emitted by the transmitting units 14 N are periodically modulated, e.g., at 10 KHz for 50 msec, off for 300 msec. Modulation of the directed beams facilitates detection thereof by the receiving subsystem 20, i.e., the receiving subsystem 20 is able to readily discriminate between modulated directed beams emitted by the transmitting subsystem 12 and any other electromagnetic radiation sources that may be active in the defined working area, e.g., television remote control units, wireless computer keyboards, microwaves, ambient radiation such as sunlight.
- any other electromagnetic radiation sources that may be active in the defined working area, e.g., television remote control units, wireless computer keyboards, microwaves, ambient radiation such as sunlight.
- the transmitting units 14 N be sequentially operated so that any transmitting unit 14 N is cycled on for a predetermined period of time and then cycled off, the next (adjacent) transmitting unit 14 N is then cycled on for the predetermined period of time and cycled off, and so forth.
- the inventors have developed two innovative ways of transmitting and processing directed beams emitted by a transmitting subsystem 12 comprised of interchangeable transmitting units 14N so that the receiving subsystem 20 can individually identify a specific interchangeable transmitting unit 14N, and, based upon such identification, establish a travel vector for the robotic device 100.
- interchangeable transmitting units 14N are operated in a predetermined manner that allows the receiving subsystem 20 to process detected directed beams to identify the directed beam having the highest signal strength, which, in turn, allows the receiving subsystem 20 to identify the interchangeable transmitting unit 14 N that emitted such directed beam. This, in turn, allows the receiving subsystem 20 to determine the orientation and, hence the travel vector, of the robotic device 100.
- the transmitting subsystem 12 is first cycled on so that all transmitting units 14 N emit directed beams for a predetermined synchronization period, as identified by reference character t SY , and then cycled off.
- the receiver subsystem 20 is operative to detect and process one or more of the directed beams emitted by the transmitting units 14 N and identify the predetermined synchronization period t SY of the transmitting subsystem 12. This identification allows the receiving subsystem 20 to synchronize operations between the transmitting subsystem 12 and the receiving subsystem 20 by initializing a timing sequence at the end of the predetermined synchronization period t SY (reference character t 0 identifies the initialization of the timing sequence in Figure 5A ).
- the transmitting subsystem 12 is further operative so that individual transmitting unit 14 N are sequentially cycled on and off at predetermined times with respect to the timing sequence initialization t 0 established by the receiving subsystem 20.
- Figure 5A which illustrates a transmitting subsystem 12 comprising four transmitting units 14 N (arbitrarily identified as the first transmitting unit 14 1 , the second transmitting unit 14 2 , the third transmitting unit 14 3 , and the fourth transmitting unit 14 4 )
- the transmitting subsystem 12 is configured and operative so that each of the transmitting units 14 1 , 14 2 , 14 3 , 14 4 is sequentially cycled on to emit a directed beam that transitions from a zero (0) signal strength to a peak signal strength to a zero (0) signal strength and then cycled off (a saw-tooth transition pattern is exemplarily illustrated in Figure 5A one skilled in the art will appreciate that other types of signal strength transition patterns can be used in the practice of the invention described herein, e.g., a ramped
- the first transmitting unit 14 1 is cycled on and transitions to a peak signal strength at time t 1 .
- the second transmitting unit 14 2 is cycled on as the directed beam from the first transmitting unit 14 1 achieves its peak signal strength at time t 1 .
- the second transmitting unit 14 2 transitions to a peak signal strength at time t 2 , at which point the first transmitting unit 14 1 has transitioned to a zero (0) signal strength and is cycled off.
- the third transmitting unit 14 3 is cycled on as the directed beam from the second transmitting unit 14 2 achieves its peak signal strength at time t 2 .
- the foregoing operating pattern is repeated for the second, third, and fourth transmitting units 14 2 , 14 3 , 14 4 , as applicable, so that at time t 3 the second transmitting unit 14 2 is cycled off, the directed beam emitted by the third transmitting unit 14 3 has achieved its peak signal strength, and the fourth transmitting unit 14 4 is cycled on; and at time t 4 the third transmitting unit 14 3 is cycled off and the directed beam emitted by the fourth transmitting unit 14 4 has achieved its peak strength.
- the transmitting subsystem 12 is operative to repeat the above-described synchronization -- sequential transmission procedure during operation of the navigational control system 10.
- interchangeable transmitting units 14N are operated in a different predetermined manner that allows the receiving subsystem 20 to process detected directed beams to identify the directed beam having the highest signal strength, which, in turn, allows the receiving subsystem 20 to identify the interchangeable transmitting unit 14 N that emitted such directed beam. This, in turn, allows the receiving subsystem 20 to determine the orientation and, hence the travel vector, of the robotic device 100.
- the transmitting subsystem 12 is first cycled on so that all transmitting units 14 N emit directed beams for a predetermined synchronization period, as identified by reference character t 12 , and then cycled off.
- the receiver subsystem 20 is operative to detect and process one or more of the directed beams emitted by the transmitting units 14 N and identify the predetermined synchronization period t 12 of the transmitting subsystem 12. This identification allows the receiving subsystem 20 to synchronize operations between the transmitting subsystem 12 and the receiving subsystem 20 by initializing a timing sequence at the end of the predetermined synchronization period t SY (reference character t 0 identifies the initialization of the timing sequence in Figure 5A ).
- the transmitting subsystem 12 is further operative so that individual transmitting unit 14 N are sequentially cycled on and off at predetermined times with respect to the timing sequence initialization t 0 established by the receiving subsystem 20.
- Figure 5C which illustrates a transmitting subsystem 12 comprising four transmitting units 14 N (arbitrarily identified as the first transmitting unit 14 1 , the second transmitting unit 14 2 , the third transmitting unit 14 3 , and the fourth transmitting unit 14 4 )
- the transmitting subsystem 12 is configured and operative so that each of the transmitting units 14 1 , 14 2 , 14 3 , 14 4 is sequentially cycled on to emit a pulsed directed beam have a predetermined pulse width P 1 , P 2 , P 3 , P 4 , respectively, at a predetermined signal strength, and then cycled off.
- the first transmitting unit 14 1 is cycled on at t 11 (where the first "1" identifies the transmitting unit number and the second "1" indicates that the transmitting unit is cycled on) and cycled off at t 12 (where the "2" indicates that the transmitting unit is cycled off).
- the second transmitting unit 14 2 is cycled on at t 21 and cycled off at t 22
- the third transmitting unit 14 3 is cycled on at t 31 and cycled off at t 32
- fourth transmitting units 14 4 is cycled on at t 41 and cycled off at t 42 .
- the transmitting subsystem 12 is operative to repeat the above-described synchronization -- sequential transmission procedure during operation of the navigational control system 12 according to the present invention.
- the transmitting units 14 N are discrete and identifiable, each transmitting unit 14 N operating at a unique operating frequency to emit a directed beam (which is preferably planar with respect to the surface of the defined working area) having a predetermined emission pattern along a specific propagation axis. These operating frequencies are also preferably modulated to facilitate detection thereof by the receiving subsystem 20 in an environment where other electromagnetic radiation sources are operating. Since each directed beam is readily and uniquely identifiable, the receiving subsystem 20 can process detected directed beams in a conventional manner to derive not only the absolute bearing and to the robotic device 100, but also the travel vector for the robotic device 100 at any particular time.
- the receiving subsystem 20 of the navigational control system 10 comprises a processing unit 22 that includes a microprocessor 24, a signal processing unit 26, a memory module 28, and a set of detection units 30 M . Additionally, the receiving subsystem 20 can also include a transmitting unit 32 for those preferred embodiments of the navigational control system 10 wherein the receiving subsystem 20 is operated or functions as the base station for the navigational control system 10.
- the memory module 28 comprises RAM 28A and ROM 28B.
- Data relating to the current operation of the robotic device 100 within the defined working area is stored in the RAM 28A.
- Such current operational data can include the grid map of cells C defining the defined working area and the position history of the robotic device 100 within the defined working area for the 'position history' embodiment of the navigational control system 10.
- Stored in the ROM 28B are one or more navigation control algorithms for the navigational control system 10, a set of one or more control signals associated with each navigation control algorithm, and a signal processing algorithm for converting signals generated by the signal processing unit 26 to one or more sets of instantaneous position parameters, i.e., a bearing, distance pair (and travel vector, if applicable).
- a set of instantaneous position parameters that define the position history of the robotic device 100, which are correlated with the grid map to identify the cells C comprising the position history.
- vigation control algorithm encompasses a set of instructions that: (a) define how the position history or instantaneous position is used by the navigational control system 10 (e.g., counting and comparing cells visited, a true-false determination for cells visited, true-false determination whether the predetermined triggering event has occurred); (b) defines the triggering event or events associated with the use of the position history or the instantaneous position; and (c) identifies the control signal(s) to be implemented when the triggering event is realized.
- the microprocessor 24 is operative to count and store the number of visits to each cell and to compute the total number of visits to cells contiguous to (neighboring) each such visited cell (cell counting). The microprocessor 24 is further operative to compare the total number of neighboring-cell visits as each cell is visited to a threshold value (see, e.g., Figure 4 wherein "C V " identifies a visited cell and "C C " identifies the eight (8) cells contiguous to the visited cell C V ).
- the microprocessor 24 is operative to cause a control signal to be communicated to the robotic device 100.
- the control signal causes the robotic device 100 to implement one or more behavioral modes specified by the control signal, e.g., a Spot Coverage pattern as described above.
- one or more cells in the stored grid map are pre-identified (i.e., prior to operating the robotic device 100) as "hot spots" in the defined working area.
- the microprocessor 24 is operative to determine whether the visited cell has been identified as a "hot spot" (true-false determination). If the microprocessor 24 determines that the visited cell C is a "hot spot" (triggering event), the microprocessor 24 is operative to cause a control signal to be communicated to the robotic device 100 via the control signal transmitting unit 32. Reception of the control signal causes the robotic device 100 to implement the prescribed conduct specified by the control signal, e.g., one or more of the basic maneuvers described above and/or a Spot Coverage pattern or Obstacle-Following behavioral mode as described above.
- navigation control algorithms for the 'position history' embodiment of the navigational control system 10 are implemented without knowledge of the travel vector of the robotic device 100, i.e., based solely upon the identification of visited cells by means of the bearing, distance parameters derived by the receiving subsystem 20.
- Another representative example of a navigation control algorithm for the 'position history' embodiment of the navigation control system 10 utilizes the travel vector as an element of the position history in issuing a control signal.
- the microprocessor 24 is operative to count and store the number of times a cell has been visited (cell counting) and further operative to compare this number to the number of times each contiguous (or neighboring) cell has been visited.
- the triggering event is a numerical differential between the number of visits to the currently-visited cell number and the number of visits to each of the neighboring-cells that identifies the neighboring cell or cells that have been least-visited as compared to the currently-visited cell.
- the triggering event would cause the receiving system 20 to issue a control signal to the robotic device 100 that causes the robotic device 100 to move from the currently-visited cell to the neighboring cell that has been visited least, e.g., by implementing one or more basic maneuvers as described herein. If two or more neighboring cells have been visited least, the control signal would cause the robotic device to move from the currently-visited cell to the least visited neighboring cell that is most compatible with the current travel vector of the robotic device 100, e.g., minimum travel distance.
- One representative navigation control algorithm for the 'instantaneous position' of the navigational control system 10 uses elapsed time (either random or predetermined) as the predetermined triggering event to cause the robotic device 10 to move to a predetermined position B in the defined working environment.
- the microprocessor 24 is operative, upon expiration of the elapsed time (the predetermined triggering event), to determine the instantaneous position (hereinafter identified as "position A") of the robotic device 100 as described herein. Since position A is an unknown variable until the predetermined triggering event is realized, the prescribed conduct, i.e., the basic maneuvers, necessary to move the robotic device 100 from position A to position B are also unknown.
- the basic maneuvers necessary to move the robotic device 100 from position A to position B are determinable since both position A and position B are known variables (in terms of their known bearing, distance parameter pairs with respect to the receiving subsystem 20).
- a determination of the basic maneuvers that will be implemented by the robotic device 100 can be accomplished by any conventional computational technique.
- Another exemplary navigation control algorithm for the 'instantaneous position' embodiment of the navigational control system 10 is a variation of the "hot spot" navigation control algorithm for the 'position history' embodiment of the navigational control system 10.
- both position A and position B are known variables and accordingly, the basic maneuver(s) to move the robotic device 100 from position A to position B are known.
- the predetermined triggering event is a TRUE determination that the instantaneous position of the robotic device 100 is equal to position A (position A may be stored in memory 28 as a "zone" - defining some arbitrary area centered about position A - rather than a single point position to increase the probability that the instantaneous position of the robotic device 100 at some time will equal position A).
- the receiving subsystem 20 comprises a set of detection units 30 M where M is an integer defining the number of individual detection units comprising the set for the navigational control system 10.
- M is an integer defining the number of individual detection units comprising the set for the navigational control system 10.
- the number and positioning of the set of detection units 30 M should be such that as much of the defined working area as possible is within the field-of-view of the receiving subsystem 20 and that the fields-of-view of at least two (but preferably more) detection units 30 M cover the same area within the defined working area.
- the receiving subsystem 20 functions as a "base station" for the system 10.
- the receiving subsystem 20 is a portable, standalone unit that is stationarily positioned within the defined working area, preferably abutting a wall bounding the defined working area (the 'wall unit' configuration).
- the receiving subsystem 20 can be positioned within the defined working area distally of the walls bounding the defined working area (the 'free-standing' configuration).
- the receiving subsystem 20 as the base station establishes and, for the 'position history' embodiment of the navigational control system 10, stores the grid map of cells representing the defined working area and represents the origin (0, 0) of the grid map of cells described above.
- the individual detection units 30 M have a known spaced-apart relationship and configured and operative to have a 180° field-of-view.
- Preferred embodiments of the wall unit configuration for the navigational control system 10 include three detection units 30 M to provide absolute bearing data to the robotic device 100.
- a minimum of two detection units 30M are required to provide the necessary signal information for the receiving subsystem 20.
- More than three detection units 30 M can be employed to increase the resolution of the receiving subsystem 20, but at an added cost for each additional detection unit 30 M and associated signal processing circuitry (see Figure 6C which illustrates the representative signal processing circuitry associated with a detection unit 30 M ).
- the individual detection units 30 M likewise spaced apart by known angular distances and configured and operative have a field-of-view greater than 180°.
- a representative embodiment of the receiving subsystem 20 operated as a free-standing base station would comprise four detection units 30 M .
- the detection units 30 M are configured and operative to detect a parameter of one or more of the directed beams emitted by the transmitting units 14 N , e.g., voltages V representing the relative signal strengths of the detected directed beam(s).
- each detection unit 30 M is configured and operative to average the detected signal strength parameter (e.g., voltage) when the detection unit 30 M detects two directed beams simultaneously.
- the receiving subsystem 20 executes a signal processing algorithm that processes the detected parameters provided by the detection units 30M, i.e., relative signal strengths of the detected beams, utilizing a conventional technique to determine the absolute bearing between the robotic device 100 and the receiving subsystem 20.
- the receiving subsystem 20 is preferably calibrated prior to use. This involves positioning the robotic device 100 at a predetermined distance from the receiving subsystem 20 and operating one (or more) of the transmitting units 14 N to emit a directed beam at the receiving subsystem 20.
- the parameter of the directed beam detected by the detection units 30 M e.g., a voltage representing the signal strength of the directed beam as detected, is correlated to the predetermined distance and used to generate a look-up table of signal strength versus distance for the defined working area.
- This look-up table is stored in the memory module 28 of the receiving subsystem 20.
- the receiving subsystem 20 uses the detected signal strengths as pointers to the stored look-up table to determine the corresponding distances (between the receiving subsystem 20 and the robotic device 100).
- the receiving subsystem 20 could be configured and operative to implement a signal processing algorithm that utilizes the known attenuation characteristics, i.e., signal strength versus distance, of the operating frequency of the directed beams emitted by the transmitting units 14 N .
- This embodiment presupposes that the transmitting units 14 N are rated and emitting directed beams of known signal strength.
- the detection units 30 M of the receiving subsystem 20 are configured to scan the set of unique operating frequencies utilized by the transmitting units 14 N .
- the receiving subsystem 20 is configured and operative to cause the detection units 30 M to sequentially scan through these frequencies during operation of the navigational control system 10.
- Figure 5B illustrates the operating characteristics of the complementary receiving subsystem 20.
- the receiving subsystem 20 is configured and operative to detect the directed beams emitted during the predetermined synchronization period t SY .
- the receiving subsystem 20 is operative to initiate the timing sequence t 0 .
- the receiving subsystem 20 is operative to detect the directed beams as described herein.
- the receiving subsystem 20 is further operative to determine the time at which the peak signal strength is detected, see reference character t peak in Figure 5B .
- the receiving subsystem 20 is further operative to correlate the peak signal strength detection time t peak with the known times at which the signal strength of the directed beam emitted by each transmitting unit 14 N reached its peak to identify the specific transmitting unit 14 N that transmitted the directed beam detected as having the peak signal strength (for the descriptive example presented in Figures 5A, 5B , the third transmitting unit 14 3 ).
- Figure 5D illustrates the operating characteristics of the complementary receiving subsystem 20.
- the receiving subsystem 20 is configured and operative to detect the directed beams emitted during the predetermined synchronization period t SY .
- the receiving subsystem 20 is operative to initiate the timing sequence t 0 .
- the receiving subsystem 20 is operative to detect the directed beams as described herein (as exemplarily illustrated by the detected signal pulses DP 1 , DP 2 , DP 3 , DP 4 in Figure 5D ).
- the receiving subsystem 20 is further operative to determine the two highest peak signal strengths of the detected directed beams, see reference characters DP 3 and DP 2 in Figure 5D , which depict the highest and next highest detected signal pulses, and the times at which the two highest strength signals were detected (t 21 and t 31 in Figure 5D ).
- the signal strength detection times allows the particular transmitting units 14 N on the robotic device 100 to be identified, i.e., transmitting units 14 3 and 14 2 in the example of Figure 5D .
- the receiving subsystem 20 is then further operative to compute the amplitude ratio of these signal pulses, e.g., DP 3 /DP 2 , and to use such computed amplitude ratio as a pointer to a look-up table that identifies the angular orientation of the identified transmitting units 14 3 , 14 2 , which in turn establishes the travel vector for the robotic device 100.
- the specific location of each individual transmitting unit 14N on the robotic device 100 is a known quantity. Based upon the identification of the transmitting unit 14N that emitted the directed beam detected by the receiving subsystem 20, the receiving subsystem 20 can execute rather straightforward geometric calculations, based upon the location of the identified transmitting unit 14N, to determine the travel vector of the robotic device 100.
- embodiments of the receiving subsystem 20 that operate as a base station further include a transmitting unit 32 (see Figure 3 ).
- the microprocessor 24 is operative to select the appropriate control signal to implement such prescribed conduct from the memory module 28.
- the microprocessor 24 is then operative to activate the transmitting unit 32 to communicate (by transmitting) the control signal to the receiver unit 16 of the robotic device 100 where the prescribed conduct defined by the control signal is implemented by means of the microprocessing unit 135.
- the robotic device 100 is described (and depicted in Figure 3 ) as being configured to include a dedicated receiver unit 16 for receiving control signals transmitted by the transmitting unit 32 of the receiving unit 20, it is preferable that the omnidirectional detector 128 (of the virtual wall detection system) be adapted to detect and process such control signals.
- the transmitting unit 32 is not required. Rather, the receiving unit 20 of the navigation control system 100 is electrically coupled to the microprocessing unit 135 (via an I/O port) of the robotic device 100 so that the receiving unit 20 can communicate control signals directly to the microprocessing unit 135.
- the receiving subsystem 20 functions as the base station, i.e., the wall unit configuration, and the transmitting subsystem 12 is integrated in combination with the robotic device 100.
- the base station i.e., the wall unit configuration
- the transmitting subsystem 12 is integrated in combination with the robotic device 100.
- One preferred embodiment that is illustrative of the features and functionality of the navigational control system 10 is exemplarily illustrated in Figures 6A-6C .
- Figure 6A depicts a robotic device 100 operating in a defined working area WA bounded by walls W.
- a virtual wall unit VWU is positioned in the only entryway to the working area WA and operative to emit a confinement beam CB that confines the robotic device 100 to operations within the working area WA.
- the transmitting subsystem 12 of the illustrated embodiment of the navigational control system 10 is integrated in combination with the robotic device 100 and comprises a set of transmitting units 14 N (eight (8) for the described embodiment such that N equals the integers 1-8) that are operative to generate a corresponding set of directed beams DB N (where N equals the integers 1-8) as illustrated in Figure 5B (only two directed beams DB 3 , DB 4 are illustrated in Figure 5B ).
- Reference characters BA 1 - BA 8 identify the propagation axes of the directed beams DB N emitted by the transmitting units 14 1 - 14 8 , respectively.
- Each transmitting unit 14 N is configured and operative to emit a directed beam DB N having a predetermined emission pattern ⁇ N centered about the corresponding beam axis BA N .
- the emission pattern ⁇ N of each directed beam DB N is approximately 100°.
- the predetermined emission pattern ⁇ N of the directed beams DB N is correlated with the number of transmitting units 14 N so that the transmitting subsystem 12 of the navigational control system 10 emulates an omnidirectional transmitting source.
- An omnidirectional transmitting source is necessary to ensure that one or more of the directed beams DB N are detected by the receiving subsystem 20 since the position and orientation of the robotic device 100 in the defined working area (e.g., in terms of its forward motion FM), with respect to the receiving station 20, is an unknown variable at any particular moment in time.
- the emission patterns ⁇ N of the directed beams DB N overlap.
- the directed beams DB 3 , DB 4 emitted by transmitting units 14 3 , 14 4 , respectively, will be detected by the detection units 30 1 , 30 2 , 30 3 of the receiving subsystem 20.
- the detection units 30 1 , 30 2 , 30 3 are operative to detect a parameter representative of the relative signal strengths of the detected beams DB 3 , DB 4 , e.g., V1, V2, V3, respectively (as disclosed above each detection unit 30N is operative to average the signal strengths when two directed beams are detected simultaneously).
- the receiving subsystem 20 is operative to implement the signal processing algorithm to compute the absolute bearing and distance between the receiving subsystem 20 and the robotic device 100.
- the receiving subsystem 20 then implements the navigation control algorithm to correlate the computed bearing and distance with one of the cells comprising the grid map of the defined working area WA stored in the memory module 28, and adds such cell to the position history of the robotic device 100 to update the position history.
- the receiving subsystem 20 is then operative under the navigation control algorithm to determine if there is a predetermined triggering event associated with this updated position history. If so, the receiving subsystem 20 is operative to select the appropriate control signal, as determined by the navigation control algorithm, and transmit such control signal to the receiver unit 16 of the robotic device 100 using the transmitting system 32 (see Figure 3 ).
- the microprocessing unit 135 of the robotic device 100 is operative in response to the reception of the control signal by means of the omnidirectional detector 128, to implement prescribed conduct, e.g., one or more of the basic maneuvers and/or behavioral modes exemplarily described herein, specified by the control signal.
- FIG. 7 An exemplary embodiment of a navigational control system 10' wherein the transmitting subsystem 12 functions as a base station and the receiving subsystem 20 is integrated in combination with the robotic device 100 is illustrated in Figure 7 .
- the transmitting subsystem 12 comprises a distributed set of transmitting units 14N positioned to abut the walls W of the defined working area.
- the transmitting subsystem 12 comprises a first transmitting unit 14 1 , a second transmitting unit 14 2 , and a third transmitting unit 14 3 positioned in abutting engagement with adjacent walls W, respectively.
- Each transmitting unit 14N comprising this distributed set is configured and operative to emit a directed beam having a predetermined emission pattern ⁇ N along a predetermined beam axis DB N (DB 1 , DB 2 , and DB 3 in Figure 7 define the predetermined beam axes for the distributed transmitting units 14 1 , 14 2 , 14 3 , respectively) at a unique operating frequency, preferably in the infrared frequency range and preferably modulated as disclosed herein.
- each transmitting unit 14 1 , 14 2 , 14 3 is configured and operative to generate a predetermined beam emission pattern ⁇ N that effectively covers the defined working area WA, i.e., ⁇ N is approximately 180° for the distributed transmission subsystem 12 depicted in Figure 7 .
- the receiving subsystem 20 for the navigational control system 10' preferably comprises a single omnidirectional detection unit 30 which may be of the type described in commonly-owned, co-pending U.S. Patent Application Serial No. 10/056,804, filed 24 January 2002 , entitled METHOD AND SYSTEM FOR ROBOT LOCALIZATION AND CONFINEMENT (the virtual wall system summarily described herein).
- the omnidirectional detection unit 30 is configured and operative to scan through the unique operating frequencies utilized by the distributed transmitting units 14 1 , 14 2 , 14 3 .
- the omnidirectional detection unit 30 is operative to detect the directed beams DB1, DB2, DB3 emitted by the distributed transmitting units 14 1 , 14 2 , 14 3 .
- the receiving subsystem is configured and operative to process the signals of the detected directed beam to determine the absolute position of the robotic device 100 within the defined working area WA. This absolute position is defined in terms of a cell of the grid map of the defined working area WA. A sequence of absolute positions, determined as described above, identifies a sequence of cells that defines the position history of the robotic device 100.
- the receiver subsystem 20 is operative as described above to utilize a navigation control algorithm to determine whether a triggering event has occurred in the position history, and if a trigger event has occurred, the receiver subsystem 20 is operative to communicate the control signal associated with the triggering event/navigation control algorithm to the robotic device 100.
- the robotic device 100 is operative, in response to the communicated control signal, to implement the prescribed conduct specified by the control signal.
- the navigational control system 10 has been described above as determining and using the instantaneous position (or a sequence of instantaneous positions) of a robotic device as a control parameter for directly altering the movement activity of the robotic device.
- the navigational control system can be used for other purposes.
- the navigational control system can be used for correcting errors in movement activity of robotic devices relying upon dead reckoning. It is therefore to be understood that, within the scope of the appended claims, the present invention may be practiced other than as specifically described herein.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Aviation & Aerospace Engineering (AREA)
- Automation & Control Theory (AREA)
- Electromagnetism (AREA)
- Control Of Position, Course, Altitude, Or Attitude Of Moving Bodies (AREA)
- Electric Vacuum Cleaner (AREA)
Description
- This nonprovisional patent application is based upon and claims priority from
U.S. provisional patent application Serial No. 60/410,480, filed 13 September 2002 - This nonprovisional patent application is related to commonly-owned, co-pending
U.S. nonprovisional patent application Serial No. 10/056,804, filed 24 January 2002 , entitled METHOD AND SYSTEM FOR ROBOT LOCALIZATION AND CONFINEMENT,U.S. nonprovisional patent application Serial No. 10/320,729, filed 16 December 2002 U.S. nonprovisional patent application Serial No. 10/167,851, filed 12 June 2002 10/453,202, filed 03 June 2003 - The present invention relates generally to mobile robotic devices, and more particularly, to a navigational control system for a robotic device operating under the direction of a navigation control algorithm that tracks the movement activity of the robotic device and utilizes movement activity to provide one or more control instructions to the robotic device, which in response thereto, implements a prescribed conduct, e.g., one or more basic maneuvers and/or behavioral patterns to increase the coverage efficiency of such robotic devices for covering a defined working area, e.g., sweeping, vacuuming a room, and/or performing residential yard maintenance.
- Robotic engineers have long worked on developing an effective method of autonomous cleaning. This has led to the development of two separate and distinct schemes for autonomous robotic devices: (1) deterministic cleaning; and (2) random cleaning.
- In deterministic cleaning, where the cleaning rate equals the coverage rate and is, therefore, a more efficient cleaning method than random-motion cleaning, the autonomous robotic device follows a defined path, e.g., a boustrophedon path that is calculated to facilitate complete cleaning coverage of a given area while eliminating redundant cleaning. Deterministic cleaning requires that the robotic device maintain precise position knowledge at all times, as well as its position history (where it has been), which, in turn, requires a sophisticated positioning system. A suitable positioning system - a positioning system suitably accurate for deterministic cleaning might rely on scanning laser ranging systems, ultrasonic transducers, a carrier phase differential GPS, or other sophisticated methods - is typically prohibitively expensive and labor intensive, requiring an involved pre-setup to accommodate the unique conditions of each area to be cleaned, e.g., room geometry, furniture locations. In addition, methods that rely on global positioning are typically incapacitated by failure of any part of the positioning system.
- One illustrative example of a highly sophisticated (and relatively expensive) robotic device for deterministic cleaning is the RoboScrub device built by Denning Mobile Robotics and Windsor Industries. The RoboScrub device employs sonar and infrared detectors, bump sensors, and a high-precision laser navigation system to define the deterministic cleaning path. The navigation system employed with the RoboScrub device requires numerous large bar code targets to be set up in various strategic positions within the area to be cleaned, and effective operation of the navigation system requires that at least four of such targets be visible simultaneously. This target accessibility requirement effectively limits the use of the RoboScrub device to large uncluttered open areas.
- Other representative deterministic robotic devices are described in
U.S. Patent Nos. 5,650,702 (Azumi ),5,548,511 (Bancroft ),5,537,017 (Feiten et al. ),5,353,224 (Lee et al. ),4,700,427 (Knepper ), and4,119,900 (Kreimnitz ). These representative deterministic robotic devices are likewise relatively expensive, require labor intensive pre-setup, and/or are effectively limited to large, uncluttered areas of simple geometric configuration (square, rectangular rooms with minimal (or no) furniture). - Due to the limitations and difficulties inherent in purely deterministic cleaning systems, some robotic devices rely on pseudo-deterministic cleaning schemes such as dead reckoning. Dead reckoning consists of continually measuring the precise rotation of each drive wheel (e.g., using optical shaft encoders) to continually calculate the current position of the robotic device, based upon a known starting point and orientation. In addition to the disadvantages of having to start cleaning operations from a fixed position with the robotic device in a specified orientation, the drive wheels of dead reckoning robotic devices are almost always subject to some degree of slippage, which leads to errors in the calculation of current position. Accordingly, dead reckoning robotic devices are generally considered unreliable for cleaning operations of any great duration - resulting in intractable system neglect, i.e., areas of the surface to be cleaned are not cleaned. Other representative examples of pseudo-deterministic robotic devices are described in
U.S. Patent Nos. 6,255,793 (Peless et al. ) and5,109,566 (Kobayashi et al. ). - A robotic device operating in random motion, under the control of one or more random-motion algorithms stored in the robotic device, represents the other basic approach to cleaning operations using autonomous robotic devices. The robotic device autonomously implement such random-motion algorithm(s) in response to internal events, e.g., signals generated by a sensor system, elapse of a time period (random or predetermined). In a typical room without obstacles, a robotic device operating under the control of a random-motion algorithm will provide acceptable cleaning coverage given enough cleaning time. Compared to a robotic device operating in a deterministic cleaning mode, a robotic device utilizing a random-motion algorithm must operate for a longer period of time to achieve acceptable cleaning coverage. To have a high confidence that a random-motion robotic device has cleaned 98% of an obstacle-free room, the random-motion robotic device must run approximately five times longer than a deterministic robotic device having similarly sized cleaning mechanisms and moving at approximately the same speed.
- However, an area to be cleaned that includes one or more randomly-situated obstacles causes a marked increase in the running time for a random-motion robotic device to effect 98% cleaning coverage. Therefore, while a random motion robotic device is a relatively inexpensive means of cleaning a defined working area as contrasted to a deterministic robotic device, the random-motion robotic device requires a significantly higher cleaning time.
- A need exists to provide a deterministic component to a random-motion robotic device to enhance the cleaning efficiency thereof to reduce the running time for the random-motion robotic cleaning to achieve a 98% cleaning coverage.
- One aim is to provide a navigational control system that enhances the cleaning efficiency of a robotic device by adding a deterministic component (in the form of a conduct prescribed by a navigation control algorithm) to the random motion of the robotic device generated by predetermined behavioral modes stored in the robotic device.
- Another aim is to provide a navigational control unit operating under a navigation control algorithm that includes a predetermined triggering event that defines when the prescribed conduct will be implemented by the robotic device.
- These are achieved by means of a navigational control system for deterministically altering movement activity of a robotic device operating in a defined working area, comprising a transmitting subsystem integrated in combination with the robotic device, the transmitting subsystem comprising means for emitting a number of directed beams, each directed beam having a predetermined emission pattern, and a receiving subsystem functioning as a base station that includes a navigation control algorithm that defines a predetermined triggering event for the navigational control system and a set of detection units positioned within the defined working area, the detection units being positioned in a known aspectual relationship with respect to one another, the set of detection units being configured and operative to detect one or more of the directed beams emitted by the transmitting system; and wherein the receiving subsystem is configured and operative to process the one or more detected directed beams under the control of the navigational control algorithm to determine whether the predetermined triggering event has occurred, and, if the predetermined triggering event has occurred transmit a control signal to the robotic device, wherein reception of the control signal by the robotic device causes the robotic device to implement a prescribed conduct that deterministically alters the movement activity of the robotic device. The invention is defined in
claim 1. - A more complete understanding of the present invention and the attendant features and advantages thereof can be had by reference to the following detailed description of the invention when considered in conjunction with the accompanying drawings wherein:
-
Figure 1 is a top-view schematic of an exemplary robotic device having particular utility for use in a navigational control system. -
Figure 2 is an exemplary hardware block diagram for the robotic device ofFigure 1 . -
Figure 3 is a schematic depiction of a navigational control system that comprises a transmitting subsystem and a receiving subsystem. -
Figure 4 illustrates a polar tessellation of a defined working area in which a robotic device is operating. -
Figure 5A illustrates the operation of one preferred embodiment of a transmitting subsystem in synchronized operation with the receiving subsystem of a preferred embodiment of the navigational control system. -
Figure 5B illustrates the operation of the receiving subsystem in synchronized operation with the transmitting subsystem ofFigure 5A . -
Figure 5C illustrates the operation of another embodiment of a transmitting subsystem in synchronized operation with the receiving subsystem of a preferred embodiment of the navigational control system. -
Figure 5D illustrates the operation of the receiving subsystem in synchronized operation with the transmitting subsystem ofFigure 5C . -
Figure 6A illustrates one preferred embodiment of a navigational control system according to the present invention wherein the transmitting subsystem in integrated in combination with the robotic device and the receiving system functions as a base station mounted against one wall of a defined working area. -
Figure 6B illustrates the set of transmitting units comprising the transmitting subsystem of the robotic device ofFigure 6A and representative directed beams having a predetermined emission patterns. -
Figure 6C is a schematic illustration of a preferred embodiment of the receiving subsystem ofFigure 6A . -
Figure 7 illustrates an exemplary embodiment of a navigational control system wherein the receiving subsystem is integrated in combination with the robotic device and the transmitting subsystem has a distributed configuration. - Referring now to the drawings wherein like reference numerals identify corresponding or similar elements throughout the several views,
Figure 1 is a top-view schematic of an exemplary preferred embodiment of arobotic device 100 having particular utility in combination with anavigational control system 10.Figure 2 is a block diagram of the hardware of therobot device 100 ofFigure 1 . - The hardware and behavioral modes (coverage behaviors for cleaning operations; escape behaviors for transitory movement patterns; and safety behaviors for emergency conditions) of the
robotic device 100, which is manufactured, distributed, and/or marketed by the iRobot Corporation of Burlington, MA under the ROOMBA trademark, are briefly described in the following paragraphs to facilitate a more complete understanding of thenavigational control system 10 of the present invention. Further details regarding the hardware and behavioral modes of therobotic device 100 can be found in commonly-owned, co-pendingU.S. nonprovisional patent application Serial No. 10/167,851, filed 12 June 2002 U.S. nonprovisional patent application Serial No. 10/320,729, filed 16 December 2002 - In the following description of the
robotic device 100, use of the terminology "forward"/"fore" refers to the primary direction of motion (forward) of the robotic device (see arrow identified by reference character "FM" inFigure 1 ). The fore/aft axis FAX of therobotic device 100 coincides with the medial diameter of therobotic device 100 that divides therobotic device 100 into generally symmetrical right and left halves, which are defined as the dominant and non-dominant sides, respectively. - The
robotic device 100 has a generally cylindrical housing infrastructure that includes achassis 102 and anouter shell 104 secured to thechassis 102 that define a structural envelope of minimal height (to facilitate movement under furniture). The hardware comprising therobotic device 100 can be generally categorized as the functional elements of a power system, a motive power system, a sensor system, a control module, a side brush assembly, or a self-adjusting cleaning head system, respectively, all of which are integrated in combination with the housing infrastructure. In addition to such categorized hardware, therobotic device 100 further includes aforward bumper 106 having a generally arcuate configuration and a nose-wheel assembly 108. - The forward bumper 106 (illustrated as a single component; alternatively, a two-segment component) is integrated in movable combination with the chassis 102 (by means of displaceable support members pairs) to extend outwardly therefrom. Whenever the
robotic device 100 impacts an obstacle (e.g., wall, furniture) during movement thereof, thebumper 106 is displaced (compressed) towards thechassis 102 and returns to its extended (operating) position when contact with the obstacle is terminated. - The nose-
wheel assembly 108 is mounted in biased combination with thechassis 102 so that the nose-wheel subassembly 108 is in a retracted position (due to the weight of the robotic device 100) during cleaning operations wherein it rotates freely over the surface being cleaned. When the nose-wheel subassembly 108 encounters a drop-off during operation (e.g., descending stairs, split-level floors), the nose-wheel assembly 108 is biased to an extended position. - The hardware of the power system, which provides the energy to power the electrically-operated hardware of the
robotic device 100, comprises a rechargeable battery pack 110 (and associated conduction lines, not shown) that is integrated in combination with thechassis 102. - The motive power system provides the means that propels the
robotic device 100 and operates the cleaning mechanisms, e.g., side brush assembly and the self-adjusting cleaning head system, during movement of therobotic device 100. The motive power system comprises left and right maindrive wheel assemblies 112L, 112R, their associated independentelectric motors 114L, 114R, andelectric motors - The main
drive wheel assemblies 112L, 112R are independently mounted in biased combination with the chassis 102 (for pivotal motion with respect thereto) at opposed ends of the transverse diameter (with respect to the fore-aft axis FAX) of therobotic device 100 and are in a retracted position (due to the weight of the robotic device 100) during operation thereof wherein the axes of rotation are approximately coplanar with the bottom of thechassis 102. If therobotic device 100 is removed from the surface being cleaned, themain wheel assemblies 112L, 112R are pivotally-biased to an extended position wherein their axes of rotation are below the bottom plane of the chassis 102 (in this extended position therechargeable battery pack 110 is automatically turned off by the control module executing one of the safety behavioral modes). - The
electric motors 114L, 114R are mechanically coupled to the maindrive wheel assemblies 112L, 112R, respectively, and independently operated by control signals generated by the control module as a response to the implementation of a behavioral mode. Independent operation of theelectric motors 114L, 114R allows themain wheel assemblies 112L, 112R to be: (1) rotated at the same speed in the same direction to propel therobotic device 100 in a straight line, forward or aft; (2) differentially rotated (including the condition wherein one wheel assembly is not rotated) to effect a variety of right and/or left turning patterns (over a spectrum of sharp to shallow turns) for therobotic device 100; and (3) rotated at the same speed in opposite directions to cause therobotic device 100 to turn in place, i.e., "spin on a dime", to provide an extensive repertoire of movement capability for therobotic device 100. - The sensor system comprises a variety of different sensor units that are operative to generate signals that control the behavioral mode operations of the
robotic device 100. The describedrobotic device 100 includesobstacle detection units 120,cliff detection units 122,wheel drop sensors 124, an obstacle-followingunit 126, a virtual wallomnidirectional detector 128, stall-sensor units 130, and mainwheel encoder units 132. - For the described embodiment, the obstacle ("bump")
detection units 120 are IR break beam sensors mounted in combination with the displaceable support member pairs of theforward bumper 106. Thesedetection units 120 are operative to generate one or more signals indicating relative displacement between one or more support member pairs whenever therobotic device 100 impacts an obstacle such that theforward bumper 106 is compressed. These signals are processed by the control module to determine an approximate point of contact with the obstacle relative to the fore-aft axis FAX of the robotic device 100 (and the behavioral mode(s) to be implemented). - The
cliff detection units 122 are mounted in combination with theforward bumper 106. Eachcliff detection unit 122 comprises an IR emitter - detector pair configured and operative to establish a focal point such that radiation emitted downwardly by the emitter is reflected from the surface being traversed and detected by the detector. If reflected radiation is not detected by the detector, i.e., a drop-off is encountered, thecliff detection unit 122 transmits a signal to the control module (which causes one or more behavioral modes to be implemented). - A
wheel drop sensor 124 such as a contact switch is integrated in combination with each of the maindrive wheel assemblies 112L, 112R and thenose wheel assembly 108 and is operative to generate a signal whenever any of the wheel assemblies is in an extended position, i.e., not in contact with the surface being traversed, (which causes the control module to implement one ore more behavioral modes). - The obstacle-following
unit 126 for the described embodiment is an IR emitter - detector pair mounted on the 'dominant' side (right hand side ofFigure 1 ) of therobotic device 100. The emitter - detector pair is similar in configuration to the cliff detection units 112, but is positioned so that the emitter emits radiation laterally from the dominant side of therobotic device 100. Theunit 126 is operative to transmit a signal to the control module whenever an obstacle is detected as a result of radiation reflected from the obstacle and detected by the detector. The control module, in response to this signal, causes one or more behavioral modes to be implemented. - A virtual wall detection system for use in conjunction with the described embodiment of the
robotic device 100 comprises anomnidirectional detector 128 mounted atop theouter shell 104 and a stand-alone transmitting unit (not shown) that transmits an axially-directed confinement beam. The stand-alone transmitting unit is positioned so that the emitted confinement beam blocks an accessway to a defined working area, thereby restricting therobotic device 100 to operations within the defined working area (e.g., in a doorway to confine therobotic device 100 within a specific room to be cleaned). Upon detection of the confinement beam, theomnidirectional detector 128 transmits a signal to the control module (which causes one or more behavioral modes to be implemented to move therobotic device 100 away from the confinement beam generated by the stand-alone transmitting unit). - A
stall sensor unit 130 is integrated in combination with eachelectric motor - An IR encoder unit 132 (see
Figure 2 ) is integrated in combination with eachmain wheel assembly 112L, 112R and operative to detect the rotation of the corresponding wheel and transmit signals corresponding thereto the control module (wheel rotation can be used to provide an estimate of distance traveled for the robotic device 100). - The control module comprises the
microprocessing unit 135 illustrated inFigure 2 that includes I/O ports connected to the sensors and controllable hardware of therobotic device 100, a microcontroller, and ROM and RAM memory. The I/O ports function as the interface between the microcontroller and the sensor units and controllable hardware, transferring signals generated by the sensor units to the microcontroller and transferring control (instruction) signals generated by the microcontroller to the controllable hardware to implement a specific behavioral mode. - The microcontroller is operative to execute instruction sets for processing sensor signals, implementing specific behavioral modes based upon such processed signals, and generating control (instruction) signals for the controllable hardware based upon implemented behavioral modes for the
robotic device 100. The cleaning coverage and control programs for therobotic device 100 are stored in the ROM of themicroprocessing unit 135, which includes the behavioral modes, sensor processing algorithms, control signal generation algorithms and a prioritization algorithm for determining which behavioral mode or modes are to be given control of therobotic device 100. The RAM of themicroprocessing unit 135 is used to store the active state of therobotic device 100, including the ID of the behavioral mode(s) under which therobotic device 100 is currently being operated and the hardware commands associated therewith. - The
side brush assembly 140 is configured and operative to entrain particulates outside the periphery of the housing infrastructure and to direct such particulates towards the self-adjusting cleaning head system. Theside brush assembly 140 provides therobotic device 100 with the capability of cleaning surfaces adjacent to base-boards when the robotic device is operated in an Obstacle-Following behavioral mode. As shown inFigure 1 , theside brush assembly 140 is preferably mounted in combination with thechassis 102 in the forward quadrant on the dominant side of therobotic device 100. - The self-adjusting
cleaning head system 145 for the describedrobotic device 100 comprises a dual-stage brush assembly and a vacuum assembly, each of which is independently powered by an electric motor (reference numeral 118 inFigure 1 actually identifies two independent electric motors -- one for the brush assembly and one for the vacuum assembly). The cleaning capability of therobotic device 100 is commonly characterized in terms of the width of the cleaning head system 145 (see reference character W inFigure 1 ). - The dual-stage brush assembly and the inlet of the vacuum assembly are integrated in combination with a deck structure, which is pivotally mounted in combination with the
chassis 102 and operatively integrated with the motor of the dual-stage brush assembly. In response to a predetermined reduction in rotational speed of the brush assembly motor, the brush assembly motor provides the motive force to pivot the deck structure with respect to thechassis 102. The pivoting deck structure provides the self adjusting capability for the cleaninghead assembly 145, which allows therobotic device 100 to readily transition between disparate surfaces during cleaning operations, e.g., carpeted surface to bare surface or vice versa, without hanging up. - The dual-stage brush assembly comprises asymmetric, counter-rotating brushes that are positioned (forward of the inlet of the vacuum assembly), configured and operative to direct particulate debris into a removable dust cartridge (not shown). The positioning, configuration, and operation of the brush assembly concomitantly directs particulate debris towards the inlet of the vacuum assembly such that particulates that are not swept up by the dual-stage brush assembly can be subsequently ingested by the vacuum assembly as a result of movement of the
robotic device 100. - Operation of the vacuum assembly independently of the self-adjustable brush assembly allows the vacuum assembly to generate and maintain a higher vacuum force using a battery-power source than would be possible if the vacuum assembly were operated in dependence with the brush assembly.
- The
robotic device 100 uses a variety of behavioral modes to effectively clean a defined working area where behavioral modes are layers of control systems that can be operated in parallel. Themicroprocessor unit 135 is operative to execute a prioritized arbitration scheme to identify and implement one or more dominant behavioral modes for any given scenario based upon inputs from the sensor system. - The behavioral modes for the described
robotic device 100 can be characterized as: (1) coverage behavioral modes; (2) escape behavioral modes; and (3) safety behavioral modes. Coverage behavioral modes are primarily designed to allow therobotic device 100 to perform its cleaning operations in an efficient and effective manner and the escape and safety behavioral modes are priority behavioral modes implemented when a signal from the sensor system indicates that normal operation of therobotic device 100 is impaired, e.g., obstacle encountered, or is likely to be impaired, e.g., drop-off detected. - Representative and illustrative coverage behavioral (cleaning) modes for the
robotic device 100 include: (1) a Spot Coverage pattern; (2) an Obstacle-Following (or Edge-Cleaning) Coverage pattern, and (3) a Room Coverage pattern. The Spot Coverage pattern causes therobotic device 100 to clean a limited area within the defined working area, e.g., a high-traffic area. In a preferred embodiment the Spot Coverage pattern is implemented by means of a spiral algorithm (but other types of self-bounded area algorithms, e.g., polygonal, can be used). The spiral algorithm, which causes outward spiraling (preferred) or inward spiraling movement of therobotic device 100, is implemented by control signals from themicroprocessing unit 135 to themain wheel assemblies 112L, 112R to change the turn radius/radii thereof as a function of time (thereby increasing/decreasing the spiral movement pattern of the robotic device 100). - The
robotic device 100 is operated in the Spot Coverage pattern for a predetermined or random period of time, for a predetermined or random distance (e.g., a maximum spiral distance) and/or until the occurrence of a specified event, e.g., activation of one or more of the obstacle detection units 120 (collectively a transition condition). Once a transition condition occurs, therobotic device 100 can implement or transition to a different behavioral mode, e.g., a Straight Line behavioral mode (in a preferred embodiment of therobotic device 100, the Straight Line behavioral mode is a low priority, default behavior that propels the robot in an approximately straight line at a preset velocity of approximately 0.306 m/s) or a Bounce behavioral mode in combination with a Straight Line behavioral mode. - If the transition condition is the result of the
robotic device 100 encountering an obstacle, therobotic device 100 can take other actions in lieu of transitioning to a different behavioral mode. Therobotic device 100 can momentarily implement a behavioral mode to avoid or escape the obstacle and resume operation under control of the spiral algorithm (i.e., continue spiraling in the same direction). Alternatively, therobotic device 100 can momentarily implement a behavioral mode to avoid or escape the obstacle and resume operation under control of the spiral algorithm (but in the opposite direction - reflective spiraling). - The Obstacle-Following Coverage pattern causes the
robotic device 100 to clean the perimeter of the defined working area, e.g., a room bounded by walls, and/or the perimeter of an obstacle (e.g., furniture) within the defined working area. Preferably therobotic device 100 utilizes obstacle-followingunit 126 to continuously maintain its position with respect to an obstacle, e.g., wall, furniture, so that the motion of therobotic device 100 causes it to travel adjacent to and concomitantly clean along the perimeter of the obstacle. Different embodiments of the obstacle-followingunit 126 can be used to implement the Obstacle-Following behavioral pattern. - In a first embodiment, the obstacle-following
unit 126 is operated to detect the presence or absence of the obstacle. In an alternative embodiment, the obstacle-followingunit 126 is operated to detect an obstacle and then maintain a predetermined distance between the obstacle and therobotic device 100. In the first embodiment, themicroprocessing unit 135 is operative, in response to signals from the obstacle-following unit, to implement small CW or CCW turns to maintain its position with respect to the obstacle. Therobotic device 100 implements a small CW when therobotic device 100 transitions from obstacle detection to non-detection (reflection to non-reflection) or to implement a small CCW turn when therobotic device 100 transitions from non-detection to detection (non-reflection to reflection). Similar turning behaviors are implemented by therobotic device 100 to maintain the predetermined distance from the obstacle. - The
robotic device 100 is operated in the Obstacle-Following behavioral mode for a predetermined or random period of time, for a predetermined or random distance (e.g., a maximum or minimum distance) and/or until the occurrence of a specified event, e.g., activation of one or more of the obstacle detection units 120 a predetermined number of times (collectively a transition condition). In certain embodiments, themicroprocessor 135 will cause the robotic device to implement an Align behavioral mode upon activation of the obstacle-detection units 120 in the Obstacle-Following behavioral mode wherein the implements a minimum angle CCW turn to align therobotic device 100 with the obstacle. - The Room Coverage pattern can be used by the
robotic device 100 to clean any defined working area that is bounded by walls, stairs, obstacles or other barriers (e.g., a virtual wall unit). A preferred embodiment for the Room Coverage pattern comprises the Random-Bounce behavioral mode in combination with the Straight Line behavioral mode. Initially, therobotic device 100 travels under control of the Straight-Line behavioral mode, i.e., straight-line algorithm (maindrive wheel assemblies 112L, 112R operating at the same rotational speed in the same direction) until an obstacle is encountered. Upon activation of one or more of theobstacle detection units 120, themicroprocessing unit 135 is operative to compute an acceptable range of new directions based upon the obstacle detection unit(s) 126 activated. Themicroprocessing unit 135 selects a new heading from within the acceptable range and implements a CW or CCW turn to achieve the new heading with minimal movement. In some embodiments, the new turn heading may be followed by forward movement to increase the cleaning efficiency of therobotic device 100. The new heading may be randomly selected across the acceptable range of headings, or based upon some statistical selection scheme, e.g., Gaussian distribution. In other embodiments of the Room Coverage behavioral mode, themicroprocessing unit 135 can be programmed to change headings randomly or at predetermined times, without input from the sensor system. - The
robotic device 100 is operated in the Room Coverage behavioral mode for a predetermined or random period of time, for a predetermined or random distance (e.g., a maximum or minimum distance) and/or until the occurrence of a specified event, e.g., activation of the obstacle-detection units 120 a predetermined number of times (collectively a transition condition). - A preferred embodiment of the
robotic device 100 includes four escape behavioral modes: a Turn behavioral mode, an Edge behavioral mode, a Wheel Drop behavioral mode, and a Slow behavioral mode. One skilled in the art will appreciate that other behavioral modes can be utilized by therobotic device 100. One or more of these behavioral modes may be implemented, for example, in response to a current rise in one of theelectric motors side brush assembly 140 or dual-stage brush assembly above a low or high stall threshold,forward bumper 106 in compressed position for determined time period, detection of a wheel-drop event. - In the Turn behavioral mode, the
robotic device 100 turns in place in a random direction, starting at higher velocity (e.g., twice normal turning velocity) and decreasing to a lower velocity (one-half normal turning velocity), i.e., small panic turns and large panic turns, respectively. Low panic turns are preferably in the range of 45° to 90°, large panic turns are preferably in the range of 90° to 270°. The Turn behavioral mode prevents therobotic device 100 from becoming stuck on room impediments, e.g., high spot in carpet, ramped lamp base, from becoming stuck under room impediments, e.g., under a sofa, or from becoming trapped in a confined area. - In the Edge behavioral mode follows the edge of an obstacle unit it has turned through a predetermined number of degrees, e.g., 60°, without activation of any of the
obstacle detection units 120, or until the robotic device has turned through a predetermined number of degrees, e.g., 170°, since initiation of the Edge behavioral mode. The Edge behavioral mode allows therobotic device 100 to move through the smallest possible openings to escape from confined areas. - In the Wheel Drop behavioral mode, the
microprocessor 135 reverses the direction of the mainwheel drive assemblies 112L, 112R momentarily, then stops them. If the activatedwheel drop sensor 124 deactivates within a predetermined time, themicroprocessor 135 then reimplements the behavioral mode that was being executed prior to the activation of thewheel drop sensor 124. - In response to certain events, e.g., activation of a
wheel drop sensor 124 or acliff detector 122, the Slow behavioral mode is implemented to slowed down therobotic device 100 for a predetermined distance and then ramped back up to its normal operating speed. - When a safety condition is detected by the sensor subsystem, e.g., a series of brush or wheel stalls that cause the corresponding electric motors to be temporarily cycled off,
wheel drop sensor 124 or acliff detection sensor 122 activated for greater that a predetermined period of time, therobotic device 100 is generally cycled to an off state. In addition, an audible alarm may be generated. - The foregoing description of behavioral modes for the
robotic device 100 are intended to be representative of the types of operating modes that can be implemented by therobotic device 100. One skilled in the art will appreciate that the behavioral modes described above can be implemented in other combinations and/or circumstances. -
Figure 3 is a schematic representation of anavigational control system 10 for use in combination with arobotic device 100 to enhance the cleaning efficiency thereof by adding a deterministic component (in the form of a control signal that remotely controls the movement of the robotic device 100) to the motion algorithms, including random motion, autonomously implemented by therobotic device 100. Thenavigational control system 10 comprises a transmittingsubsystem 12 and a receivingsubsystem 20 operating under the direction of a navigation control algorithm. The navigation control algorithm includes a definition of a predetermined triggering event. The specific features and characteristics of the transmittingsubsystem 12 and the receivingsubsystem 20 depend upon whether the particular subsystem is integrated in combination with therobotic device 100 or functions as a "base station" for thenavigational control system 10. - Broadly described, the
navigational control system 10 is operative, under the direction of the navigation control algorithm, to monitor the movement activity of therobotic device 100 within the defined working area. The monitored movement activity is defined in terms of the "position history" of therobotic device 100 as described in further detail below. In a preferred embodiment, the monitored movement activity is further defined in terms of the "instantaneous position" of therobotic device 100 as defined in further detail below. - The predetermined triggering event is a specific occurrence or condition in the movement activity of the
robotic device 100. Upon the realization of the predetermined triggering event, thenavigational control system 10 is operative to generate and communicate a control signal to therobotic device 100. In response to the control signal, therobotic device 100 is operative to implement or execute a conduct prescribed by the control signal, i.e., the prescribed conduct. This prescribed conduct represents a deterministic component of the movement activity of therobotic device 100. - In the preferred embodiment of the
navigational control system 10 based upon position history, thesystem 10 is configured and operative to create a "tessellation" of any defined working area where therobotic device 100 is to be operated, e.g., a room to be cleaned. Tessellate is used herein in the sense that the defined working area is segmented into a set of individual cells, which may or may not be of equal size. For example,Figure 4 exemplarily illustrates the polar tessellation of a defined working area into a set of individual cells C (reference characters BST identify the "base station") of unequal size. The position of each cell C (in terms of its center) is identified in terms of polar coordinates (r, θ) referenced to the base station BST as the origin (0, 0). A grid map of the cells C comprising the defined working area is stored in memory of thenavigation control system 10. One skilled in the art will appreciate that other coordinate systems, e.g., a planar Cartesian coordinate system, can be used by thenavigational control system 10 to define the position of individual cells C within the predetermined working area. - Preferably, the
navigational control system 10 is operative to define the size the individual cells C so that the length and width dimensions of an individual cell C are no larger than one-half the width (W) of thecleaning head system 145 of the robotic device 100 (seeFigure 1 and corresponding discussion above). - The
navigational control system 10 is operative to generate a position history of therobotic device 100 within the defined working area in terms of such individual cells C (to minimize the memory requirements for storage of position history). The position history comprises a set of discrete, instantaneous positions (in terms of individual cells C) of therobotic device 100 over a time interval where the time interval is a variable that depends upon the "triggering condition" of the navigation control algorithm implemented by thenavigational control system 10. - Each discrete instantaneous position of the
robotic device 100 is determined by operating the transmittingsubsystem 12 to emit a set of directional beams and operating the receivingsubsystem 20 to detect one or more of such directional beams and process a signal parameter of the detected beam(s) to determine an absolute bearing parameter and a distance parameter between the transmittingsubsystem 12 and the receivingsubsystem 20 at a point in time. Each pair of bearing, distance parameters establishes a discrete instantaneous position for therobotic device 100. For the preferred 'position history' embodiment, thenavigational control system 10 is operative to correlate each discrete instantaneous position to one individual cell C of the grid map. A set of bearing and position pairs, i.e., a set of instantaneous positions, over a time interval defines a set of cells C, which are identified in the receivingsubsystem 20 as the position history of therobotic device 100 for the time interval. - In the preferred embodiment of the
navigational control system 10 based upon the instantaneous position, thesystem 10 processes each discrete instantaneous position as it is established, under the control of the navigation control algorithm, to determine whether such discrete instantaneous position is the predetermined triggering event defined by the navigation control algorithm. - In an advanced embodiment of the
navigational control system 10, thesystem 10 is additionally configured and operative to determine a travel vector (indicating the direction of motion of therobotic device 100 within an individual cell C or at the discrete instantaneous position) at each point in time. These travel vectors may be stored in memory in conjunction with the corresponding cells C as a component of the position history of therobotic device 100. - The
navigational control system 10 according to the present invention is further operative, under direction of the navigational control algorithm, to generate and communicate a control signal to therobotic device 100 whenever thenavigational control system 100 realizes the predetermined triggering event. In response to any such control signal, therobotic device 100 is configured and operative to initiate a prescribed conduct. The prescribed conduct comprises the deterministic component added to the random motion movement activity of therobotic device 100 by means of thenavigational control system 10 according to the present invention. - In one preferred embodiment of the invention, the prescribed conduct of the
robotic device 100 comprises one or more basic maneuvers such as CW and CCW turns, forward or aft (straight line) movement, slow down, speed up, and stop. The CW and/or CCW turns can be implemented using the turning techniques of therobotic device 100 described above, and the turn angles can be, for example, over a 360° spectrum at predetermined intervals, e.g., 5° or 10°. Alternatively, or in addition to, the CW and/or CCW turns can be to a specified azimuthal heading (referenced to the base station as the origin) where thenavigational control system 10 is configured and operative so that the travel vector is a determinable variable. Of these basic maneuvers, forward (straight line) movement is typically the default maneuver that therobotic device 100 automatically reverts to (implements) once one or more of the other basic maneuvers has been completed. - In another preferred embodiment of the invention, the prescribed conduct of the
robotic device 100 comprises one or more of the behavioral modes described herein. In yet a further preferred embodiment, the prescribed conduct of therobotic device 100 comprises a combination of the basic maneuvers and the behavioral modes described herein. - The transmitting
subsystem 12 is operative to transmit a number of directed beams having a predetermined emission pattern along a specific propagation axis. Preferably, the directed beams are planar, i.e., substantially parallel to the surface of the defined working area. - In preferred embodiments of the
navigational control system 10, the transmittingsubsystem 12 is integrated in combination with therobotic device 100. The transmittingsubsystem 12 is configured and operative to functionally emulate an omnidirectional transmission source with respect to the defined working area, i.e., by emitting a plurality of directed beams that cover the defined working area. For these preferred embodiments, therobotic device 100 further includes a receiver unit 16 (seeFigure 3 ) configured and operative to receive control signals from the receiving subsystem 20 (see discussion below regarding the transmittingunit 32 of the receiving subsystem 20). While thereceiver unit 16 is depicted as a dedicated receiving unit for the control signals, it is preferable that the omnidirectional detector 128 (of the virtual wall detection system) described above be adapted to detect and process such control signals. - In one preferred embodiment, the transmitting
subsystem 12 comprises a conventional mechanical sweeping transmitter, e.g., a laser, that is integrated in combination with a high point of the housing infrastructure of therobotic device 100 so that none of the structural features of therobotic device 100 interfere with the operation thereof. The mechanical sweeping transmitter is configured and operative to emit the plurality of directed beams while concomitantly redirecting (mechanically sweeping) the transmitting element so that each directed beam has a different propagation axis. Other features and characteristics of the mechanical sweeping transmitter are described below in terms ofindividual transmitting units 14N for ease of description. - Another preferred embodiment of the transmitting
subsystem 12 comprises a set of transmittingunits 14N, where N is an integer defining the number of individual transmitting units comprising the set for thenavigational control system 10, that are integrated in combination with therobotic device 100 about the periphery of its housing infrastructure. Each transmitting unit 14N is configured and operative to emit a directed beam having a predetermined emission pattern along a specific propagation axis. Preferably, the transmittingsubsystem 12 is configured and operative so that the emitted directed beams are planar. - In a basic embodiment of the transmitting
subsystem 12, the transmittingunits 14N are fungible/interchangeable, each operating to emit a directed beam at a common operating frequency. Preferably, the common operating frequency for the transmittingunits 14N lies in the infrared range, i.e., about 750 nm to about 1.4 x 104 nm, preferably about 880 nm to about 980 nm, although one skilled in the art will appreciate that other wavelengths, e.g., in the radio frequency range, microwave frequency range, can be used in the practice of thenavigational control system 10. - Preferably, the common operating frequency directed beams emitted by the transmitting
units 14N are periodically modulated, e.g., at 10 KHz for 50 msec, off for 300 msec. Modulation of the directed beams facilitates detection thereof by the receivingsubsystem 20, i.e., the receivingsubsystem 20 is able to readily discriminate between modulated directed beams emitted by the transmittingsubsystem 12 and any other electromagnetic radiation sources that may be active in the defined working area, e.g., television remote control units, wireless computer keyboards, microwaves, ambient radiation such as sunlight. For the basic embodiment, it is also preferable that the transmittingunits 14N be sequentially operated so that any transmittingunit 14N is cycled on for a predetermined period of time and then cycled off, the next (adjacent) transmittingunit 14N is then cycled on for the predetermined period of time and cycled off, and so forth. Operating the transmittingsubsystem 12 in the foregoing manner, i.e., modulation of the directed beam,cycling transmitting units 14N on/off sequentially, minimizes the power requirements of the transmittingsubsystem 12 and reduces spurious noise/collateral energy that could adversely impact the functioning of thenavigational control system 10. - Ordinarily, a
navigational control system 10 employing the basic embodiment of the transmittingsubsystem 12, i.e., all transmittingunits 14N are interchangeableemitting directed beams at a common operating frequency, cannot be used to determine travel vectors for therobotic device 100 because the receivingsubsystem 20 cannot differentiate between directed beams emitted by the transmittingunits 14N and therefore cannot identify any particular transmittingunit 14N. However, the inventors have developed two innovative ways of transmitting and processing directed beams emitted by a transmittingsubsystem 12 comprised of interchangeable transmitting units 14N so that the receivingsubsystem 20 can individually identify a specific interchangeable transmitting unit 14N, and, based upon such identification, establish a travel vector for therobotic device 100. - Accordingly, in one enhanced version of the basic embodiment of the transmitting
subsystem 12, interchangeable transmitting units 14N are operated in a predetermined manner that allows the receivingsubsystem 20 to process detected directed beams to identify the directed beam having the highest signal strength, which, in turn, allows the receivingsubsystem 20 to identify theinterchangeable transmitting unit 14N that emitted such directed beam. This, in turn, allows the receivingsubsystem 20 to determine the orientation and, hence the travel vector, of therobotic device 100. - Referring to
Figure 5A , the transmittingsubsystem 12 is first cycled on so that all transmittingunits 14N emit directed beams for a predetermined synchronization period, as identified by reference character tSY, and then cycled off. Thereceiver subsystem 20 is operative to detect and process one or more of the directed beams emitted by the transmittingunits 14N and identify the predetermined synchronization period tSY of the transmittingsubsystem 12. This identification allows the receivingsubsystem 20 to synchronize operations between the transmittingsubsystem 12 and the receivingsubsystem 20 by initializing a timing sequence at the end of the predetermined synchronization period tSY (reference character t0 identifies the initialization of the timing sequence inFigure 5A ). - The transmitting
subsystem 12 is further operative so thatindividual transmitting unit 14N are sequentially cycled on and off at predetermined times with respect to the timing sequence initialization t0 established by the receivingsubsystem 20. For example, with respect toFigure 5A , which illustrates a transmittingsubsystem 12 comprising four transmitting units 14N (arbitrarily identified as thefirst transmitting unit 141, thesecond transmitting unit 142, thethird transmitting unit 143, and the fourth transmitting unit 144), the transmittingsubsystem 12 is configured and operative so that each of the transmittingunits Figure 5A one skilled in the art will appreciate that other types of signal strength transition patterns can be used in the practice of the invention described herein, e.g., a ramped signal strength). - That is, the
first transmitting unit 141 is cycled on and transitions to a peak signal strength at time t1. Thesecond transmitting unit 142 is cycled on as the directed beam from thefirst transmitting unit 141 achieves its peak signal strength at time t1. Thesecond transmitting unit 142 transitions to a peak signal strength at time t2, at which point thefirst transmitting unit 141 has transitioned to a zero (0) signal strength and is cycled off. Thethird transmitting unit 143 is cycled on as the directed beam from thesecond transmitting unit 142 achieves its peak signal strength at time t2. The foregoing operating pattern is repeated for the second, third, and fourth transmittingunits second transmitting unit 142 is cycled off, the directed beam emitted by thethird transmitting unit 143 has achieved its peak signal strength, and thefourth transmitting unit 144 is cycled on; and at time t4 thethird transmitting unit 143 is cycled off and the directed beam emitted by thefourth transmitting unit 144 has achieved its peak strength. The transmittingsubsystem 12 is operative to repeat the above-described synchronization -- sequential transmission procedure during operation of thenavigational control system 10. - In another enhanced version of the basic embodiment of the transmitting
subsystem 12, interchangeable transmitting units 14N are operated in a different predetermined manner that allows the receivingsubsystem 20 to process detected directed beams to identify the directed beam having the highest signal strength, which, in turn, allows the receivingsubsystem 20 to identify theinterchangeable transmitting unit 14N that emitted such directed beam. This, in turn, allows the receivingsubsystem 20 to determine the orientation and, hence the travel vector, of therobotic device 100. - Referring to
Figure 5C , the transmittingsubsystem 12 is first cycled on so that all transmittingunits 14N emit directed beams for a predetermined synchronization period, as identified by reference character t12, and then cycled off. Thereceiver subsystem 20 is operative to detect and process one or more of the directed beams emitted by the transmittingunits 14N and identify the predetermined synchronization period t12 of the transmittingsubsystem 12. This identification allows the receivingsubsystem 20 to synchronize operations between the transmittingsubsystem 12 and the receivingsubsystem 20 by initializing a timing sequence at the end of the predetermined synchronization period tSY (reference character t0 identifies the initialization of the timing sequence inFigure 5A ). - The transmitting
subsystem 12 is further operative so thatindividual transmitting unit 14N are sequentially cycled on and off at predetermined times with respect to the timing sequence initialization t0 established by the receivingsubsystem 20. For example, with respect toFigure 5C , which illustrates a transmittingsubsystem 12 comprising four transmitting units 14N (arbitrarily identified as thefirst transmitting unit 141, thesecond transmitting unit 142, thethird transmitting unit 143, and the fourth transmitting unit 144), the transmittingsubsystem 12 is configured and operative so that each of the transmittingunits - That is, the
first transmitting unit 141 is cycled on at t11 (where the first "1" identifies the transmitting unit number and the second "1" indicates that the transmitting unit is cycled on) and cycled off at t12 (where the "2" indicates that the transmitting unit is cycled off). In a similar manner, thesecond transmitting unit 142 is cycled on at t21 and cycled off at t22, thethird transmitting unit 143 is cycled on at t31 and cycled off at t32, and fourth transmittingunits 144 is cycled on at t41 and cycled off at t42. The transmittingsubsystem 12 is operative to repeat the above-described synchronization -- sequential transmission procedure during operation of thenavigational control system 12 according to the present invention. - In a more sophisticated embodiment of the transmitting
subsystem 12, the transmittingunits 14N are discrete and identifiable, each transmittingunit 14N operating at a unique operating frequency to emit a directed beam (which is preferably planar with respect to the surface of the defined working area) having a predetermined emission pattern along a specific propagation axis. These operating frequencies are also preferably modulated to facilitate detection thereof by the receivingsubsystem 20 in an environment where other electromagnetic radiation sources are operating. Since each directed beam is readily and uniquely identifiable, the receivingsubsystem 20 can process detected directed beams in a conventional manner to derive not only the absolute bearing and to therobotic device 100, but also the travel vector for therobotic device 100 at any particular time. - The receiving
subsystem 20 of thenavigational control system 10 according to the present invention comprises aprocessing unit 22 that includes amicroprocessor 24, asignal processing unit 26, a memory module 28, and a set of detection units 30M. Additionally, the receivingsubsystem 20 can also include a transmittingunit 32 for those preferred embodiments of thenavigational control system 10 wherein the receivingsubsystem 20 is operated or functions as the base station for thenavigational control system 10. - The memory module 28 comprises
RAM 28A andROM 28B. Data relating to the current operation of therobotic device 100 within the defined working area is stored in theRAM 28A. Such current operational data can include the grid map of cells C defining the defined working area and the position history of therobotic device 100 within the defined working area for the 'position history' embodiment of thenavigational control system 10. Stored in theROM 28B are one or more navigation control algorithms for thenavigational control system 10, a set of one or more control signals associated with each navigation control algorithm, and a signal processing algorithm for converting signals generated by thesignal processing unit 26 to one or more sets of instantaneous position parameters, i.e., a bearing, distance pair (and travel vector, if applicable). For the 'position history' embodiment of thesystem 10, a set of instantaneous position parameters that define the position history of therobotic device 100, which are correlated with the grid map to identify the cells C comprising the position history. - The terminology "navigation control algorithm" as used herein encompasses a set of instructions that: (a) define how the position history or instantaneous position is used by the navigational control system 10 (e.g., counting and comparing cells visited, a true-false determination for cells visited, true-false determination whether the predetermined triggering event has occurred); (b) defines the triggering event or events associated with the use of the position history or the instantaneous position; and (c) identifies the control signal(s) to be implemented when the triggering event is realized. For example, in one representative navigation control algorithm for the 'position history' embodiment of the
navigational control system 10 according to the present invention, themicroprocessor 24 is operative to count and store the number of visits to each cell and to compute the total number of visits to cells contiguous to (neighboring) each such visited cell (cell counting). Themicroprocessor 24 is further operative to compare the total number of neighboring-cell visits as each cell is visited to a threshold value (see, e.g.,Figure 4 wherein "CV" identifies a visited cell and "CC" identifies the eight (8) cells contiguous to the visited cell CV). If the total number of neighboring-visits (e.g., fifteen (15) in the example ofFigure 4 ) for any visited cell is below the threshold number (the triggering event), themicroprocessor 24 is operative to cause a control signal to be communicated to therobotic device 100. The control signal causes therobotic device 100 to implement one or more behavioral modes specified by the control signal, e.g., a Spot Coverage pattern as described above. - In another representative navigation control algorithm for the 'position history' embodiment of the
navigational control system 10, one or more cells in the stored grid map are pre-identified (i.e., prior to operating the robotic device 100) as "hot spots" in the defined working area. As therobotic device 100 visits any particular cell C, themicroprocessor 24 is operative to determine whether the visited cell has been identified as a "hot spot" (true-false determination). If themicroprocessor 24 determines that the visited cell C is a "hot spot" (triggering event), themicroprocessor 24 is operative to cause a control signal to be communicated to therobotic device 100 via the controlsignal transmitting unit 32. Reception of the control signal causes therobotic device 100 to implement the prescribed conduct specified by the control signal, e.g., one or more of the basic maneuvers described above and/or a Spot Coverage pattern or Obstacle-Following behavioral mode as described above. - The foregoing representative examples of navigation control algorithms for the 'position history' embodiment of the
navigational control system 10 are implemented without knowledge of the travel vector of therobotic device 100, i.e., based solely upon the identification of visited cells by means of the bearing, distance parameters derived by the receivingsubsystem 20. Another representative example of a navigation control algorithm for the 'position history' embodiment of thenavigation control system 10 utilizes the travel vector as an element of the position history in issuing a control signal. - The
microprocessor 24 is operative to count and store the number of times a cell has been visited (cell counting) and further operative to compare this number to the number of times each contiguous (or neighboring) cell has been visited. For this navigation control algorithm, the triggering event is a numerical differential between the number of visits to the currently-visited cell number and the number of visits to each of the neighboring-cells that identifies the neighboring cell or cells that have been least-visited as compared to the currently-visited cell. The triggering event would cause the receivingsystem 20 to issue a control signal to therobotic device 100 that causes therobotic device 100 to move from the currently-visited cell to the neighboring cell that has been visited least, e.g., by implementing one or more basic maneuvers as described herein. If two or more neighboring cells have been visited least, the control signal would cause the robotic device to move from the currently-visited cell to the least visited neighboring cell that is most compatible with the current travel vector of therobotic device 100, e.g., minimum travel distance. - Using
Figure 4 as an example wherein "CV" identifies the currently-visited cell and "CC" identifies the eight (8) cells contiguous to or neighboring the currently-visited cell CV, the neighboring cells CC that have been visited a single time are the least-visited neighboring cells Cc. If the current travel vector for therobotic device 100 is indicated by the reference characters TV, the control signal would cause therobotic device 100 to continue moving in a straight line, i.e., the move forward basic maneuver (or the Straight-Line behavioral mode) would be executed by the robotic device 100 (if therobotic device 100 was currently operating in some other behavioral mode). - One representative navigation control algorithm for the 'instantaneous position' of the
navigational control system 10 uses elapsed time (either random or predetermined) as the predetermined triggering event to cause therobotic device 10 to move to a predetermined position B in the defined working environment. Themicroprocessor 24 is operative, upon expiration of the elapsed time (the predetermined triggering event), to determine the instantaneous position (hereinafter identified as "position A") of therobotic device 100 as described herein. Since position A is an unknown variable until the predetermined triggering event is realized, the prescribed conduct, i.e., the basic maneuvers, necessary to move therobotic device 100 from position A to position B are also unknown. Once position A has been determined by thenavigational control system 10, the basic maneuvers necessary to move therobotic device 100 from position A to position B are determinable since both position A and position B are known variables (in terms of their known bearing, distance parameter pairs with respect to the receiving subsystem 20). A determination of the basic maneuvers that will be implemented by therobotic device 100 can be accomplished by any conventional computational technique. - Another exemplary navigation control algorithm for the 'instantaneous position' embodiment of the
navigational control system 10 is a variation of the "hot spot" navigation control algorithm for the 'position history' embodiment of thenavigational control system 10. In this illustrative embodiment, both position A and position B are known variables and accordingly, the basic maneuver(s) to move therobotic device 100 from position A to position B are known. In this example, the predetermined triggering event is a TRUE determination that the instantaneous position of therobotic device 100 is equal to position A (position A may be stored in memory 28 as a "zone" - defining some arbitrary area centered about position A - rather than a single point position to increase the probability that the instantaneous position of therobotic device 100 at some time will equal position A). - The receiving
subsystem 20 comprises a set of detection units 30M where M is an integer defining the number of individual detection units comprising the set for thenavigational control system 10. The number and positioning of the set of detection units 30M should be such that as much of the defined working area as possible is within the field-of-view of the receivingsubsystem 20 and that the fields-of-view of at least two (but preferably more) detection units 30M cover the same area within the defined working area. - In preferred embodiments of the
navigational control system 10, the receivingsubsystem 20 functions as a "base station" for thesystem 10. In this functional role, the receivingsubsystem 20 is a portable, standalone unit that is stationarily positioned within the defined working area, preferably abutting a wall bounding the defined working area (the 'wall unit' configuration). Alternatively, the receivingsubsystem 20 can be positioned within the defined working area distally of the walls bounding the defined working area (the 'free-standing' configuration). The receivingsubsystem 20 as the base station establishes and, for the 'position history' embodiment of thenavigational control system 10, stores the grid map of cells representing the defined working area and represents the origin (0, 0) of the grid map of cells described above. - For those embodiments where the receiving
subsystem 20 is operated as a wall unit configuration, the individual detection units 30M have a known spaced-apart relationship and configured and operative to have a 180° field-of-view. For example,Figure 2 illustrates an embodiment of the receivingsubsystem 20 comprising two detection units 30M (M=2) spaced apart by a known angular distance "φ".Figure 6C illustrates another embodiment of the receivingsubsystem 20 comprising three detection units 30M (M=3), i.e., 3012, 3023, 3013, having known angular separations identified by "φ12", "φ23", and "φ13", respectively. Preferred embodiments of the wall unit configuration for thenavigational control system 10 include three detection units 30M to provide absolute bearing data to therobotic device 100. A minimum of twodetection units 30M are required to provide the necessary signal information for the receivingsubsystem 20. More than three detection units 30M can be employed to increase the resolution of the receivingsubsystem 20, but at an added cost for each additional detection unit 30M and associated signal processing circuitry (seeFigure 6C which illustrates the representative signal processing circuitry associated with a detection unit 30M). - For those embodiments where the receiving
subsystem 20 is operated as a free-standing configuration, the individual detection units 30M likewise spaced apart by known angular distances and configured and operative have a field-of-view greater than 180°. A representative embodiment of the receivingsubsystem 20 operated as a free-standing base station would comprise four detection units 30M. - The detection units 30M are configured and operative to detect a parameter of one or more of the directed beams emitted by the transmitting
units 14N, e.g., voltages V representing the relative signal strengths of the detected directed beam(s). In a preferred embodiment, each detection unit 30M is configured and operative to average the detected signal strength parameter (e.g., voltage) when the detection unit 30M detects two directed beams simultaneously. The receivingsubsystem 20 executes a signal processing algorithm that processes the detected parameters provided by thedetection units 30M, i.e., relative signal strengths of the detected beams, utilizing a conventional technique to determine the absolute bearing between therobotic device 100 and the receivingsubsystem 20. - To provide the distance determination capability for the receiving
subsystem 20, the receivingsubsystem 20 is preferably calibrated prior to use. This involves positioning therobotic device 100 at a predetermined distance from the receivingsubsystem 20 and operating one (or more) of the transmittingunits 14N to emit a directed beam at the receivingsubsystem 20. The parameter of the directed beam detected by the detection units 30M, e.g., a voltage representing the signal strength of the directed beam as detected, is correlated to the predetermined distance and used to generate a look-up table of signal strength versus distance for the defined working area. This look-up table is stored in the memory module 28 of the receivingsubsystem 20. As the signal strengths of directed beams are detected during operation of thenavigational control system 10, the receivingsubsystem 20 uses the detected signal strengths as pointers to the stored look-up table to determine the corresponding distances (between the receivingsubsystem 20 and the robotic device 100). - Alternatively, the receiving
subsystem 20 could be configured and operative to implement a signal processing algorithm that utilizes the known attenuation characteristics, i.e., signal strength versus distance, of the operating frequency of the directed beams emitted by the transmittingunits 14N. This embodiment presupposes that the transmittingunits 14N are rated and emitting directed beams of known signal strength. - For the sophisticated embodiment of the
navigational control system 10 described above wherein theindividual transmitting units 14N of the transmittingsubsystem 12 are operated at a unique operating frequency, the detection units 30M of the receivingsubsystem 20 are configured to scan the set of unique operating frequencies utilized by the transmittingunits 14N. The receivingsubsystem 20 is configured and operative to cause the detection units 30M to sequentially scan through these frequencies during operation of thenavigational control system 10. - For the innovative embodiment of the transmitting
subsystem 12 described above in connection withFigure 5A, Figure 5B illustrates the operating characteristics of thecomplementary receiving subsystem 20. The receivingsubsystem 20 is configured and operative to detect the directed beams emitted during the predetermined synchronization period tSY. At the end of the predetermined synchronization period tSY, the receivingsubsystem 20 is operative to initiate the timing sequence t0. The receivingsubsystem 20 is operative to detect the directed beams as described herein. However, the receivingsubsystem 20 is further operative to determine the time at which the peak signal strength is detected, see reference character tpeak inFigure 5B . The receivingsubsystem 20 is further operative to correlate the peak signal strength detection time tpeak with the known times at which the signal strength of the directed beam emitted by each transmittingunit 14N reached its peak to identify thespecific transmitting unit 14N that transmitted the directed beam detected as having the peak signal strength (for the descriptive example presented inFigures 5A, 5B , the third transmitting unit 143). - For the innovative embodiment of the transmitting
subsystem 12 described above in connection withFigure 5C, Figure 5D illustrates the operating characteristics of thecomplementary receiving subsystem 20. The receivingsubsystem 20 is configured and operative to detect the directed beams emitted during the predetermined synchronization period tSY. At the end of the predetermined synchronization period tSY, the receivingsubsystem 20 is operative to initiate the timing sequence t0. The receivingsubsystem 20 is operative to detect the directed beams as described herein (as exemplarily illustrated by the detected signal pulses DP1, DP2, DP3, DP4 inFigure 5D ). However, the receivingsubsystem 20 is further operative to determine the two highest peak signal strengths of the detected directed beams, see reference characters DP3 and DP2 inFigure 5D , which depict the highest and next highest detected signal pulses, and the times at which the two highest strength signals were detected (t21 and t31 inFigure 5D ). - The signal strength detection times allows the
particular transmitting units 14N on therobotic device 100 to be identified, i.e., transmittingunits Figure 5D . The receivingsubsystem 20 is then further operative to compute the amplitude ratio of these signal pulses, e.g., DP3/DP2, and to use such computed amplitude ratio as a pointer to a look-up table that identifies the angular orientation of the identified transmittingunits robotic device 100. - Even though the transmitting
units 14N mounted in combination with therobotic device 100 are interchangeable, the specific location of each individual transmitting unit 14N on therobotic device 100 is a known quantity. Based upon the identification of the transmitting unit 14N that emitted the directed beam detected by the receivingsubsystem 20, the receivingsubsystem 20 can execute rather straightforward geometric calculations, based upon the location of the identified transmitting unit 14N, to determine the travel vector of therobotic device 100. - When the receiving
subsystem 20 functions as the base station, a means is required to communicate the control signal to the robotic device. Accordingly, embodiments of the receivingsubsystem 20 that operate as a base station further include a transmitting unit 32 (seeFigure 3 ). Once the navigation control algorithm implemented by themicroprocessor 24 has determined the prescribed conduct to be implemented by therobotic device 10, themicroprocessor 24 is operative to select the appropriate control signal to implement such prescribed conduct from the memory module 28. Themicroprocessor 24 is then operative to activate the transmittingunit 32 to communicate (by transmitting) the control signal to thereceiver unit 16 of therobotic device 100 where the prescribed conduct defined by the control signal is implemented by means of themicroprocessing unit 135. - While the
robotic device 100 is described (and depicted inFigure 3 ) as being configured to include adedicated receiver unit 16 for receiving control signals transmitted by the transmittingunit 32 of the receivingunit 20, it is preferable that the omnidirectional detector 128 (of the virtual wall detection system) be adapted to detect and process such control signals. For those embodiments of thenavigational control system 10 wherein the receivingunit 20 is integrated in combination with therobotic device 10, the transmittingunit 32 is not required. Rather, the receivingunit 20 of thenavigation control system 100 is electrically coupled to the microprocessing unit 135 (via an I/O port) of therobotic device 100 so that the receivingunit 20 can communicate control signals directly to themicroprocessing unit 135. - As disclosed above, in preferred embodiments of the
navigational control system 10, the receivingsubsystem 20 functions as the base station, i.e., the wall unit configuration, and the transmittingsubsystem 12 is integrated in combination with therobotic device 100. One preferred embodiment that is illustrative of the features and functionality of thenavigational control system 10 is exemplarily illustrated inFigures 6A-6C . -
Figure 6A depicts arobotic device 100 operating in a defined working area WA bounded by walls W. A virtual wall unit VWU is positioned in the only entryway to the working area WA and operative to emit a confinement beam CB that confines therobotic device 100 to operations within the working area WA. - The transmitting
subsystem 12 of the illustrated embodiment of thenavigational control system 10 is integrated in combination with therobotic device 100 and comprises a set of transmitting units 14N (eight (8) for the described embodiment such that N equals the integers 1-8) that are operative to generate a corresponding set of directed beams DBN (where N equals the integers 1-8) as illustrated inFigure 5B (only two directed beams DB3, DB4 are illustrated inFigure 5B ). Reference characters BA1 - BA8 identify the propagation axes of the directed beams DBN emitted by the transmitting units 141 - 148, respectively. Each transmittingunit 14N is configured and operative to emit a directed beam DBN having a predetermined emission pattern θN centered about the corresponding beam axis BAN. For the illustrated embodiment, the emission pattern θN of each directed beam DBN is approximately 100°. - Preferably, the predetermined emission pattern θN of the directed beams DBN is correlated with the number of transmitting
units 14N so that the transmittingsubsystem 12 of thenavigational control system 10 emulates an omnidirectional transmitting source. An omnidirectional transmitting source is necessary to ensure that one or more of the directed beams DBN are detected by the receivingsubsystem 20 since the position and orientation of therobotic device 100 in the defined working area (e.g., in terms of its forward motion FM), with respect to the receivingstation 20, is an unknown variable at any particular moment in time. Preferably the emission patterns θN of the directed beams DBN overlap. - As an examination of
Figures 6A, 6B (and in particularFigure 6B ) shows, the directed beams DB3, DB4 emitted by transmittingunits subsystem 20. The detection units 301, 302, 303 are operative to detect a parameter representative of the relative signal strengths of the detected beams DB3, DB4, e.g., V1, V2, V3, respectively (as disclosed above each detection unit 30N is operative to average the signal strengths when two directed beams are detected simultaneously). - The receiving
subsystem 20 is operative to implement the signal processing algorithm to compute the absolute bearing and distance between the receivingsubsystem 20 and therobotic device 100. The receivingsubsystem 20 then implements the navigation control algorithm to correlate the computed bearing and distance with one of the cells comprising the grid map of the defined working area WA stored in the memory module 28, and adds such cell to the position history of therobotic device 100 to update the position history. The receivingsubsystem 20 is then operative under the navigation control algorithm to determine if there is a predetermined triggering event associated with this updated position history. If so, the receivingsubsystem 20 is operative to select the appropriate control signal, as determined by the navigation control algorithm, and transmit such control signal to thereceiver unit 16 of therobotic device 100 using the transmitting system 32 (seeFigure 3 ). Themicroprocessing unit 135 of therobotic device 100, is operative in response to the reception of the control signal by means of theomnidirectional detector 128, to implement prescribed conduct, e.g., one or more of the basic maneuvers and/or behavioral modes exemplarily described herein, specified by the control signal. - An exemplary embodiment of a navigational control system 10' wherein the transmitting
subsystem 12 functions as a base station and the receivingsubsystem 20 is integrated in combination with therobotic device 100 is illustrated inFigure 7 . The transmittingsubsystem 12 comprises a distributed set of transmitting units 14N positioned to abut the walls W of the defined working area. As illustrated inFigure 7 , the transmittingsubsystem 12 comprises afirst transmitting unit 141, asecond transmitting unit 142, and athird transmitting unit 143 positioned in abutting engagement with adjacent walls W, respectively. - Each transmitting unit 14N comprising this distributed set is configured and operative to emit a directed beam having a predetermined emission pattern θN along a predetermined beam axis DBN (DB1, DB2, and DB3 in
Figure 7 define the predetermined beam axes for the distributed transmittingunits unit transmission subsystem 12 depicted inFigure 7 . - The receiving
subsystem 20 for the navigational control system 10' preferably comprises a single omnidirectional detection unit 30 which may be of the type described in commonly-owned, co-pendingU.S. Patent Application Serial No. 10/056,804, filed 24 January 2002 , entitled METHOD AND SYSTEM FOR ROBOT LOCALIZATION AND CONFINEMENT (the virtual wall system summarily described herein). The omnidirectional detection unit 30 is configured and operative to scan through the unique operating frequencies utilized by the distributed transmittingunits - The omnidirectional detection unit 30 is operative to detect the directed beams DB1, DB2, DB3 emitted by the distributed transmitting
units robotic device 100 within the defined working area WA. This absolute position is defined in terms of a cell of the grid map of the defined working area WA. A sequence of absolute positions, determined as described above, identifies a sequence of cells that defines the position history of therobotic device 100. - The
receiver subsystem 20 is operative as described above to utilize a navigation control algorithm to determine whether a triggering event has occurred in the position history, and if a trigger event has occurred, thereceiver subsystem 20 is operative to communicate the control signal associated with the triggering event/navigation control algorithm to therobotic device 100. Therobotic device 100 is operative, in response to the communicated control signal, to implement the prescribed conduct specified by the control signal. - A variety of modifications and variations of the present invention are possible in light of the above teachings. The
navigational control system 10 has been described above as determining and using the instantaneous position (or a sequence of instantaneous positions) of a robotic device as a control parameter for directly altering the movement activity of the robotic device. One skilled in the art will appreciate that the navigational control system can be used for other purposes. For example, the navigational control system can be used for correcting errors in movement activity of robotic devices relying upon dead reckoning. It is therefore to be understood that, within the scope of the appended claims, the present invention may be practiced other than as specifically described herein.
Claims (15)
- A robotic cleaning device (100) including:a motive power system comprising left and right main drive wheel assemblies (112L, 112R) and associated independent electric motors (114L, 114R); anda navigational control system (10) configured to monitor movement activity of the robotic cleaning device (100) within a defined working area, wherein the monitored movement activity is defined in terms of a position history of the robotic cleaning device (100), the navigational control system (10) comprising:a transmitting subsystem (12) integrated in combination with the robotic device,characterised in that the transmitting subsystem (12) comprises a sweeping laser that is integrated in combination with an upper portion of a housing infrastructure of the robotic cleaning device (100) so that none of the structural features of the robotic cleaning device (100) interfere with the operation thereof,wherein the laser is operative to functionally emulate an omnidirectional transmission source by sweeping through a 360° azimuth and emitting a plurality of directed beams that cover the defined working area substantially parallel to the surface of the defined working area.
- The robotic cleaning device (100) of claim 1, wherein the sweeping laser is a mechanical sweeping laser.
- The robotic cleaning device (100) of claim 1 or claim 2, wherein the upper portion is a high point of the housing infrastructure
- The robotic cleaning device (100) of any one of claims 1 to 3, wherein the housing infrastructure defines a structural envelope facilitating movement under furniture.
- The robotic cleaning device (100) of any one of the preceding claims, wherein the monitored movement activity is further defined in terms of the instantaneous position of the robotic cleaning device (100).
- The robotic cleaning device of any one of the preceding claims 1 to 5, wherein the navigational control system (10) is configured to create a tessellation of the defined working area by segmenting the defined working area into a set of individual cells.
- The robotic cleaning device of any one of claims 1 to 6, further comprising a wheel drop sensor (124) integrated in combination with each of the main drive wheel assemblies (112L, 112R) and operative to generate a signal whenever any of the main drive wheel assemblies (112L, 112R) is in an extended position.
- The robotic cleaning device of any one of claims 1 to 7, further comprising an obstacle-following unit (126) comprising an IR emitter-detector pair configured to transmit a signal to a control module if an obstacle is detected as a result of radiation reflected from the obstacle and detected by the detector.
- The robotic cleaning device of any one of claims 1 to 8, further comprising respective stall-sensor units (130) integrated in combination with each respective independent electric motor (114L, 114R) and operative to transmit a signal to a control module when a change in current is detected in the respective independent electric motor (114L, 114R).
- The robotic cleaning device of any one of claims 1 to 9, further comprising at least one cliff detection unit (122) comprising an IR emitter-detector pair configured to transmit a signal to a control module when a drop-off is encountered.
- The robotic cleaning device of any one of claims 1 to 10, wherein the laser is configured to cycle on/off sequentially.
- The robotic cleaning device of any one of claims 1 to 11, wherein the navigational control system is configured to deterministically alter movement activity of the robotic device operating in a defined working area.
- The robotic cleaning device of any one of claims 1 to 12, wherein the navigational control system is configured to implement one or more cleaning modes for the robotic cleaning device (100) including a spot coverage pattern, an obstacle-following coverage pattern and a room coverage pattern.
- The robotic cleaning device of claim 13, wherein the obstacle-following coverage pattern causes the robotic cleaning device (100) to clean a perimeter of a room bounded by walls.
- The robotic cleaning device of any one of claims 1 to 14, wherein one or more cells in a stored grid map are pre-identified as hot spots in the defined working area, wherein as the robotic cleaning device (100) visits any particular cell, the robotic cleaning device (100) is operative to determine whether the visited cell has been identified as a hot spot, and, if the robotic cleaning device (100) determines that the visited cell as a hot spot, the robotic cleaning device (100) is operative to implement a prescribed conduct, optionally a spot coverage pattern or an obstacle-following behavioral mode.
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US41048002P | 2002-09-13 | 2002-09-13 | |
EP03752285.1A EP1547361B1 (en) | 2002-09-13 | 2003-09-12 | A navigational control system for a robotic device |
EP16155717.8A EP3043544B1 (en) | 2002-09-13 | 2003-09-12 | A navigational control system for a robotic device |
PCT/US2003/028624 WO2004025947A2 (en) | 2002-09-13 | 2003-09-12 | A navigational control system for a robotic device |
Related Parent Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP03752285.1A Division EP1547361B1 (en) | 2002-09-13 | 2003-09-12 | A navigational control system for a robotic device |
EP16155717.8A Division EP3043544B1 (en) | 2002-09-13 | 2003-09-12 | A navigational control system for a robotic device |
Publications (2)
Publication Number | Publication Date |
---|---|
EP3361716A1 EP3361716A1 (en) | 2018-08-15 |
EP3361716B1 true EP3361716B1 (en) | 2021-04-07 |
Family
ID=31994142
Family Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP16155717.8A Expired - Lifetime EP3043544B1 (en) | 2002-09-13 | 2003-09-12 | A navigational control system for a robotic device |
EP03752285.1A Expired - Lifetime EP1547361B1 (en) | 2002-09-13 | 2003-09-12 | A navigational control system for a robotic device |
EP18164153.1A Expired - Lifetime EP3361716B1 (en) | 2002-09-13 | 2003-09-12 | Robotic cleaning device |
Family Applications Before (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP16155717.8A Expired - Lifetime EP3043544B1 (en) | 2002-09-13 | 2003-09-12 | A navigational control system for a robotic device |
EP03752285.1A Expired - Lifetime EP1547361B1 (en) | 2002-09-13 | 2003-09-12 | A navigational control system for a robotic device |
Country Status (5)
Country | Link |
---|---|
US (2) | US7024278B2 (en) |
EP (3) | EP3043544B1 (en) |
AU (1) | AU2003270581A1 (en) |
ES (1) | ES2674568T3 (en) |
WO (1) | WO2004025947A2 (en) |
Families Citing this family (276)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8412377B2 (en) * | 2000-01-24 | 2013-04-02 | Irobot Corporation | Obstacle following sensor scheme for a mobile robot |
US8788092B2 (en) | 2000-01-24 | 2014-07-22 | Irobot Corporation | Obstacle following sensor scheme for a mobile robot |
US6956348B2 (en) | 2004-01-28 | 2005-10-18 | Irobot Corporation | Debris sensor for cleaning apparatus |
US6690134B1 (en) | 2001-01-24 | 2004-02-10 | Irobot Corporation | Method and system for robot localization and confinement |
US7571511B2 (en) * | 2002-01-03 | 2009-08-11 | Irobot Corporation | Autonomous floor-cleaning robot |
US7663333B2 (en) | 2001-06-12 | 2010-02-16 | Irobot Corporation | Method and system for multi-mode coverage for an autonomous robot |
US8396592B2 (en) | 2001-06-12 | 2013-03-12 | Irobot Corporation | Method and system for multi-mode coverage for an autonomous robot |
US9128486B2 (en) | 2002-01-24 | 2015-09-08 | Irobot Corporation | Navigational control system for a robotic device |
AU2003218917A1 (en) * | 2002-01-31 | 2003-09-02 | Solar And Robotics Sa | Improvement to a method for controlling an autonomous mobile robot et related device |
US6925357B2 (en) | 2002-07-25 | 2005-08-02 | Intouch Health, Inc. | Medical tele-robotic system |
US20040162637A1 (en) | 2002-07-25 | 2004-08-19 | Yulun Wang | Medical tele-robotic system with a master remote station with an arbitrator |
ES2674568T3 (en) * | 2002-09-13 | 2018-07-02 | Irobot Corporation | Navigation control system for a robotic device |
US8386081B2 (en) | 2002-09-13 | 2013-02-26 | Irobot Corporation | Navigational control system for a robotic device |
US8428778B2 (en) | 2002-09-13 | 2013-04-23 | Irobot Corporation | Navigational control system for a robotic device |
US7346428B1 (en) * | 2002-11-22 | 2008-03-18 | Bissell Homecare, Inc. | Robotic sweeper cleaner with dusting pad |
KR100561855B1 (en) | 2002-12-30 | 2006-03-16 | 삼성전자주식회사 | Robot localization system |
KR100492588B1 (en) * | 2003-01-23 | 2005-06-03 | 엘지전자 주식회사 | Position information recognition apparatus for automatic running vacuum cleaner |
US20050010331A1 (en) * | 2003-03-14 | 2005-01-13 | Taylor Charles E. | Robot vacuum with floor type modes |
KR100486737B1 (en) * | 2003-04-08 | 2005-05-03 | 삼성전자주식회사 | Method and apparatus for generating and tracing cleaning trajectory for home cleaning robot |
US7170252B2 (en) * | 2003-07-23 | 2007-01-30 | Renesas Technology Corporation | Location aware automata |
AU2004202834B2 (en) * | 2003-07-24 | 2006-02-23 | Samsung Gwangju Electronics Co., Ltd. | Robot Cleaner |
US7813836B2 (en) | 2003-12-09 | 2010-10-12 | Intouch Technologies, Inc. | Protocol for a remotely controlled videoconferencing robot |
KR100565227B1 (en) * | 2003-12-22 | 2006-03-30 | 엘지전자 주식회사 | Position recognition apparatus and method for mobile robot |
US7332890B2 (en) | 2004-01-21 | 2008-02-19 | Irobot Corporation | Autonomous robot auto-docking and energy management systems and methods |
AU2004316156B2 (en) | 2004-01-21 | 2010-09-02 | Irobot Corporation | Method of docking an autonomous robot |
US20050204438A1 (en) | 2004-02-26 | 2005-09-15 | Yulun Wang | Graphical interface for a remote presence system |
DE102004014273A1 (en) * | 2004-03-22 | 2005-10-13 | BSH Bosch und Siemens Hausgeräte GmbH | Surface machining system |
JP2007530978A (en) | 2004-03-29 | 2007-11-01 | エヴォリューション ロボティクス インコーポレイテッド | Position estimation method and apparatus using reflected light source |
JP4477924B2 (en) * | 2004-03-31 | 2010-06-09 | 本田技研工業株式会社 | Mobile robot external detection device |
TWI262777B (en) * | 2004-04-21 | 2006-10-01 | Jason Yan | Robotic vacuum cleaner |
KR100544480B1 (en) * | 2004-05-12 | 2006-01-24 | 삼성광주전자 주식회사 | Automatic cleaning apparatus |
EP1776623B1 (en) | 2004-06-24 | 2011-12-07 | iRobot Corporation | Remote control scheduler and method for autonomous robotic device |
US8972052B2 (en) | 2004-07-07 | 2015-03-03 | Irobot Corporation | Celestial navigation system for an autonomous vehicle |
US7706917B1 (en) | 2004-07-07 | 2010-04-27 | Irobot Corporation | Celestial navigation system for an autonomous robot |
US8077963B2 (en) | 2004-07-13 | 2011-12-13 | Yulun Wang | Mobile robot with a head-based movement mapping scheme |
KR100641113B1 (en) * | 2004-07-30 | 2006-11-02 | 엘지전자 주식회사 | Mobile robot and his moving control method |
US7840308B2 (en) * | 2004-09-10 | 2010-11-23 | Honda Motor Co., Ltd. | Robot device control based on environment and position of a movable robot |
KR100656701B1 (en) * | 2004-10-27 | 2006-12-13 | 삼성광주전자 주식회사 | Robot cleaner system and Method for return to external charge apparatus |
WO2006058125A2 (en) * | 2004-11-23 | 2006-06-01 | S. C. Johnson & Son, Inc. | Device and methods of providing air purification in combination with cleaning of surfaces |
KR20060059006A (en) * | 2004-11-26 | 2006-06-01 | 삼성전자주식회사 | Method and apparatus of self-propelled mobile unit with obstacle avoidance during wall-following |
US7222000B2 (en) * | 2005-01-18 | 2007-05-22 | Intouch Technologies, Inc. | Mobile videoconferencing platform with automatic shut-off features |
US7620476B2 (en) | 2005-02-18 | 2009-11-17 | Irobot Corporation | Autonomous surface cleaning robot for dry cleaning |
AU2006214016B2 (en) | 2005-02-18 | 2011-11-10 | Irobot Corporation | Autonomous surface cleaning robot for wet and dry cleaning |
US8392021B2 (en) | 2005-02-18 | 2013-03-05 | Irobot Corporation | Autonomous surface cleaning robot for wet cleaning |
KR100653165B1 (en) * | 2005-02-25 | 2006-12-05 | 삼성광주전자 주식회사 | Robot cleaner and method of control thereof |
EP1870010A4 (en) * | 2005-03-25 | 2008-07-23 | Toshiba Tec Kk | Electric cleaner |
US8930023B2 (en) | 2009-11-06 | 2015-01-06 | Irobot Corporation | Localization by learning of wave-signal distributions |
US7353034B2 (en) | 2005-04-04 | 2008-04-01 | X One, Inc. | Location sharing and tracking using mobile phones or other wireless devices |
KR100624387B1 (en) * | 2005-04-25 | 2006-09-20 | 엘지전자 주식회사 | Robot system which can designate a movement area |
TWI278731B (en) * | 2005-05-09 | 2007-04-11 | Infinite Electronics Inc | Self-propelled apparatus for virtual wall system |
US20100222926A1 (en) | 2005-05-09 | 2010-09-02 | Ting-Yin Chiu | Virtual wall system |
US8364366B2 (en) * | 2005-06-24 | 2013-01-29 | Deere & Company | System and method for providing a safety zone associated with a vehicle |
EP1906807B1 (en) * | 2005-07-08 | 2012-01-25 | Ab Electrolux | Robotic cleaning device |
CN101253299B (en) * | 2005-07-20 | 2013-03-06 | 欧普帝姆斯特许股份公司 | Re-design of hospital operating room |
US7456596B2 (en) * | 2005-08-19 | 2008-11-25 | Cisco Technology, Inc. | Automatic radio site survey using a robot |
US9198728B2 (en) | 2005-09-30 | 2015-12-01 | Intouch Technologies, Inc. | Multi-camera mobile teleconferencing platform |
CA2625895C (en) * | 2005-10-14 | 2016-05-17 | Aethon, Inc. | Robotic ordering and delivery apparatuses, systems and methods |
DE112006002894B4 (en) * | 2005-10-21 | 2021-11-11 | Deere & Company | Networked multipurpose robotic vehicle |
TW200719559A (en) * | 2005-11-07 | 2007-05-16 | Su-Miao Liu | Charging device and method for self-propelling-type mouse |
ES2413862T3 (en) | 2005-12-02 | 2013-07-17 | Irobot Corporation | Modular robot |
EP2544066B1 (en) | 2005-12-02 | 2018-10-17 | iRobot Corporation | Robot system |
ATE442617T1 (en) * | 2005-12-02 | 2009-09-15 | Irobot Corp | ROBOT NAVIGATION SYSTEM WITH AUTONOMOUS COVER |
ES2706727T3 (en) | 2005-12-02 | 2019-04-01 | Irobot Corp | Robot system |
JP4875102B2 (en) * | 2005-12-02 | 2012-02-15 | アイロボット コーポレイション | Coverage robot mobility |
EP2816434A3 (en) | 2005-12-02 | 2015-01-28 | iRobot Corporation | Autonomous coverage robot |
ES2378138T3 (en) | 2005-12-02 | 2012-04-09 | Irobot Corporation | Robot covering mobility |
US20070150094A1 (en) * | 2005-12-23 | 2007-06-28 | Qingfeng Huang | System and method for planning and indirectly guiding robotic actions based on external factor tracking and analysis |
US8577538B2 (en) * | 2006-07-14 | 2013-11-05 | Irobot Corporation | Method and system for controlling a remote vehicle |
US8311845B2 (en) * | 2006-02-07 | 2012-11-13 | Groupon, Inc. | Pay-for-visit advertising based on visits to physical locations |
ES2707155T3 (en) * | 2006-03-17 | 2019-04-02 | Irobot Corp | Robot confinement |
WO2007117606A2 (en) * | 2006-04-07 | 2007-10-18 | Pelago, Inc. | Proximity-based user interaction |
US9100454B2 (en) | 2006-04-07 | 2015-08-04 | Groupon, Inc. | Method and system for enabling the creation and maintenance of proximity-related user groups |
US10827315B2 (en) | 2006-04-07 | 2020-11-03 | Groupon, Inc. | Facilitating user interactions based on proximity |
US8326469B2 (en) * | 2006-07-14 | 2012-12-04 | Irobot Corporation | Autonomous behaviors for a remote vehicle |
US8108092B2 (en) * | 2006-07-14 | 2012-01-31 | Irobot Corporation | Autonomous behaviors for a remote vehicle |
US20070266520A1 (en) * | 2006-05-16 | 2007-11-22 | Bosses Mark D | Vacuum cleaner brush roll control device |
ES2693223T3 (en) | 2006-05-19 | 2018-12-10 | Irobot Corporation | Removal of waste from cleaning robots |
TWI293555B (en) * | 2006-05-23 | 2008-02-21 | Ind Tech Res Inst | Omni-directional robot cleaner |
US8417383B2 (en) | 2006-05-31 | 2013-04-09 | Irobot Corporation | Detecting robot stasis |
US20070291128A1 (en) * | 2006-06-15 | 2007-12-20 | Yulun Wang | Mobile teleconferencing system that projects an image provided by a mobile robot |
US8849679B2 (en) | 2006-06-15 | 2014-09-30 | Intouch Technologies, Inc. | Remote controlled robot system that provides medical images |
US7974738B2 (en) | 2006-07-05 | 2011-07-05 | Battelle Energy Alliance, Llc | Robotics virtual rail system and method |
US7668621B2 (en) * | 2006-07-05 | 2010-02-23 | The United States Of America As Represented By The United States Department Of Energy | Robotic guarded motion system and method |
US8271132B2 (en) * | 2008-03-13 | 2012-09-18 | Battelle Energy Alliance, Llc | System and method for seamless task-directed autonomy for robots |
US7620477B2 (en) * | 2006-07-05 | 2009-11-17 | Battelle Energy Alliance, Llc | Robotic intelligence kernel |
US7801644B2 (en) * | 2006-07-05 | 2010-09-21 | Battelle Energy Alliance, Llc | Generic robot architecture |
US7587260B2 (en) * | 2006-07-05 | 2009-09-08 | Battelle Energy Alliance, Llc | Autonomous navigation system and method |
US8965578B2 (en) | 2006-07-05 | 2015-02-24 | Battelle Energy Alliance, Llc | Real time explosive hazard information sensing, processing, and communication for autonomous operation |
US8355818B2 (en) | 2009-09-03 | 2013-01-15 | Battelle Energy Alliance, Llc | Robots, systems, and methods for hazard evaluation and visualization |
US8073564B2 (en) * | 2006-07-05 | 2011-12-06 | Battelle Energy Alliance, Llc | Multi-robot control interface |
US7584020B2 (en) * | 2006-07-05 | 2009-09-01 | Battelle Energy Alliance, Llc | Occupancy change detection system and method |
US8843244B2 (en) * | 2006-10-06 | 2014-09-23 | Irobot Corporation | Autonomous behaviors for a remove vehicle |
US20100066587A1 (en) * | 2006-07-14 | 2010-03-18 | Brian Masao Yamauchi | Method and System for Controlling a Remote Vehicle |
US20080092324A1 (en) * | 2006-10-18 | 2008-04-24 | Guten Electronics Industrial Co., Ltd. | Dust-collecting auxiliary device for vacuum cleaner |
EP1913856B1 (en) * | 2006-10-20 | 2011-07-27 | Wessel-Werk GmbH | Compact suction cleaning device for autonomous cleaning of floor coverings |
KR100767120B1 (en) * | 2006-11-28 | 2007-10-17 | 삼성광주전자 주식회사 | Robot cleaner and control method thereof |
KR100815570B1 (en) * | 2006-12-06 | 2008-03-20 | 삼성광주전자 주식회사 | System for robot cleaner and control methord thereof |
WO2008118532A2 (en) | 2007-01-30 | 2008-10-02 | Fiore Industries, Inc. | Method and apparatus for remotely disabling vechicles |
US8265793B2 (en) | 2007-03-20 | 2012-09-11 | Irobot Corporation | Mobile robot for telecommunication |
US20080229885A1 (en) * | 2007-03-22 | 2008-09-25 | Mah Pat Y | Jar opener |
US20080229528A1 (en) * | 2007-03-23 | 2008-09-25 | Gooten Innolife Corporation | Floor-cleaning device |
WO2008134595A1 (en) * | 2007-04-27 | 2008-11-06 | Pelago, Inc. | Determining locations of interest based on user visits |
KR101505380B1 (en) | 2007-05-09 | 2015-03-23 | 아이로보트 코퍼레이션 | Surface treatment robot |
US9160783B2 (en) * | 2007-05-09 | 2015-10-13 | Intouch Technologies, Inc. | Robot system that operates through a network firewall |
US8255092B2 (en) | 2007-05-14 | 2012-08-28 | Irobot Corporation | Autonomous behaviors for a remote vehicle |
DE102007036229A1 (en) * | 2007-08-02 | 2009-02-05 | BSH Bosch und Siemens Hausgeräte GmbH | Mobile device i.e. self-propelled accumulator-operating dust collecting robot, introducing method for stationary device such as accumulator-loading station, involves introducing mobile device in accordance with motion-sequence |
DE102007036230B4 (en) * | 2007-08-02 | 2011-03-17 | BSH Bosch und Siemens Hausgeräte GmbH | A method and system for determining the position of a mobile device with respect to a stationary device, in particular an accumulator-powered dust collection robot with respect to a rechargeable battery charger, stationary device and mobile device |
US8116910B2 (en) * | 2007-08-23 | 2012-02-14 | Intouch Technologies, Inc. | Telepresence robot with a printer |
WO2009057918A1 (en) * | 2007-10-30 | 2009-05-07 | Lg Electronics Inc. | Detecting apparatus of robot cleaner and controlling method of robot cleaner |
US20090140926A1 (en) * | 2007-12-04 | 2009-06-04 | Elden Douglas Traster | System and method for localization utilizing dynamically deployable beacons |
CN101970187A (en) * | 2008-01-28 | 2011-02-09 | 塞格瑞德公司 | Methods for repurposing temporal-spatial information collected by service robots |
US8838268B2 (en) * | 2008-01-28 | 2014-09-16 | Seegrid Corporation | Service robot and method of operating same |
US8755936B2 (en) * | 2008-01-28 | 2014-06-17 | Seegrid Corporation | Distributed multi-robot system |
JP5548139B2 (en) * | 2008-01-28 | 2014-07-16 | シーグリッド コーポレーション | How to interact with a robot that maintains a facility in real time or near real time |
US10875182B2 (en) | 2008-03-20 | 2020-12-29 | Teladoc Health, Inc. | Remote presence system mounted to operating room hardware |
US8179418B2 (en) | 2008-04-14 | 2012-05-15 | Intouch Technologies, Inc. | Robotic based health care system |
US8170241B2 (en) * | 2008-04-17 | 2012-05-01 | Intouch Technologies, Inc. | Mobile tele-presence system with a microphone system |
DE102008028931A1 (en) * | 2008-06-18 | 2009-12-24 | BSH Bosch und Siemens Hausgeräte GmbH | Robot i.e. dust collecting robot, drive movement controlling method, involves stopping drive movement of robot during determination of coordinate values based on comparison of coordinate values of virtual partial region and/or virtual wall |
US9193065B2 (en) | 2008-07-10 | 2015-11-24 | Intouch Technologies, Inc. | Docking system for a tele-presence robot |
US9842192B2 (en) | 2008-07-11 | 2017-12-12 | Intouch Technologies, Inc. | Tele-presence robot system with multi-cast features |
TWM348676U (en) * | 2008-07-22 | 2009-01-11 | Iner Aec Executive Yuan | Environmental survey robot |
US8301325B2 (en) * | 2008-07-28 | 2012-10-30 | Precise Path Robotics, Inc. | System and method for autonomous vehicle localization |
US8340819B2 (en) | 2008-09-18 | 2012-12-25 | Intouch Technologies, Inc. | Mobile videoconferencing robot system with network adaptive driving |
CN101685160A (en) * | 2008-09-26 | 2010-03-31 | 鸿富锦精密工业(深圳)有限公司 | Environment sensor for moving device |
US8996165B2 (en) * | 2008-10-21 | 2015-03-31 | Intouch Technologies, Inc. | Telepresence robot with a camera boom |
US8463435B2 (en) * | 2008-11-25 | 2013-06-11 | Intouch Technologies, Inc. | Server connectivity control for tele-presence robot |
US9138891B2 (en) | 2008-11-25 | 2015-09-22 | Intouch Technologies, Inc. | Server connectivity control for tele-presence robot |
US8849680B2 (en) | 2009-01-29 | 2014-09-30 | Intouch Technologies, Inc. | Documentation through a remote presence robot |
EP3629051A1 (en) | 2009-03-02 | 2020-04-01 | Diversey, Inc. | Hygiene monitoring and management system and method |
US8897920B2 (en) | 2009-04-17 | 2014-11-25 | Intouch Technologies, Inc. | Tele-presence robot system with software modularity, projector and laser pointer |
US8146695B1 (en) * | 2009-04-28 | 2012-04-03 | Ernie Lance Ramshur | Automated garbage receptacle conveyance system |
US20100292884A1 (en) * | 2009-05-12 | 2010-11-18 | Rogelio Manfred Neumann | Device for Influencing Navigation of an Autonomous Vehicle |
GB0909148D0 (en) | 2009-05-28 | 2009-07-01 | F Robotics Acquisitions Ltd | Localisation system |
US8774970B2 (en) | 2009-06-11 | 2014-07-08 | S.C. Johnson & Son, Inc. | Trainable multi-mode floor cleaning device |
US8364309B1 (en) | 2009-07-14 | 2013-01-29 | Bailey Bendrix L | User-assisted robot navigation system |
US11399153B2 (en) | 2009-08-26 | 2022-07-26 | Teladoc Health, Inc. | Portable telepresence apparatus |
US8384755B2 (en) | 2009-08-26 | 2013-02-26 | Intouch Technologies, Inc. | Portable remote presence robot |
KR101406186B1 (en) * | 2009-11-18 | 2014-06-13 | 삼성전자주식회사 | Control method for a robot cleaner |
US11154981B2 (en) | 2010-02-04 | 2021-10-26 | Teladoc Health, Inc. | Robot user interface for telepresence robot system |
CN108378771B (en) | 2010-02-16 | 2021-06-11 | 艾罗伯特公司 | Vacuum cleaner brush |
US9631219B2 (en) * | 2010-02-24 | 2017-04-25 | Nalco Company | Metabolic rate indicator for cellular populations |
US8670017B2 (en) | 2010-03-04 | 2014-03-11 | Intouch Technologies, Inc. | Remote presence system including a cart that supports a robot face and an overhead camera |
TWM387645U (en) * | 2010-03-30 | 2010-09-01 | cheng-xiang Yan | Sensory escape device of automatic vacuum cleaner |
US8496737B2 (en) * | 2010-05-06 | 2013-07-30 | Moneual Inc. | Movable air purification robot system |
US9014848B2 (en) | 2010-05-20 | 2015-04-21 | Irobot Corporation | Mobile robot system |
US8935005B2 (en) | 2010-05-20 | 2015-01-13 | Irobot Corporation | Operating a mobile robot |
US8918213B2 (en) | 2010-05-20 | 2014-12-23 | Irobot Corporation | Mobile human interface robot |
US10343283B2 (en) | 2010-05-24 | 2019-07-09 | Intouch Technologies, Inc. | Telepresence robot system that can be accessed by a cellular phone |
TWI424296B (en) * | 2010-05-25 | 2014-01-21 | Micro Star Int Co Ltd | Guidance device and operation system utilizing the same |
US10808882B2 (en) | 2010-05-26 | 2020-10-20 | Intouch Technologies, Inc. | Tele-robotic system with a robot face placed on a chair |
US20120065829A1 (en) * | 2010-09-15 | 2012-03-15 | Jyh-Cheng Yu | Wall-following Moving Device |
KR20120049533A (en) * | 2010-11-09 | 2012-05-17 | 삼성전자주식회사 | Robot system and method for controlling the same |
US9264664B2 (en) | 2010-12-03 | 2016-02-16 | Intouch Technologies, Inc. | Systems and methods for dynamic bandwidth allocation |
DE102011003064A1 (en) * | 2010-12-29 | 2012-07-05 | Robert Bosch Gmbh | Method for processing a surface by means of a robot vehicle |
CN103534659B (en) | 2010-12-30 | 2017-04-05 | 美国iRobot公司 | Cover robot navigation |
US8930019B2 (en) | 2010-12-30 | 2015-01-06 | Irobot Corporation | Mobile human interface robot |
EP3026423B1 (en) | 2010-12-30 | 2024-05-08 | iRobot Corporation | Debris monitoring in a dust bin |
US12093036B2 (en) | 2011-01-21 | 2024-09-17 | Teladoc Health, Inc. | Telerobotic system with a dual application screen presentation |
CN103459099B (en) | 2011-01-28 | 2015-08-26 | 英塔茨科技公司 | Mutually exchange with a moveable tele-robotic |
US9323250B2 (en) | 2011-01-28 | 2016-04-26 | Intouch Technologies, Inc. | Time-dependent navigation of telepresence robots |
BR112013026178A2 (en) | 2011-04-11 | 2019-10-01 | Crown Equipment Ltd | method and system for coordinating route planning |
US10769739B2 (en) | 2011-04-25 | 2020-09-08 | Intouch Technologies, Inc. | Systems and methods for management of information among medical providers and facilities |
EP2703925B1 (en) * | 2011-04-28 | 2021-02-24 | Positec Power Tools (Suzhou) Co., Ltd | Automatic working system, automatic walking device and steering method thereof |
US20140139616A1 (en) | 2012-01-27 | 2014-05-22 | Intouch Technologies, Inc. | Enhanced Diagnostics for a Telepresence Robot |
US9098611B2 (en) | 2012-11-26 | 2015-08-04 | Intouch Technologies, Inc. | Enhanced video interaction for a user interface of a telepresence network |
KR101931365B1 (en) * | 2011-08-22 | 2018-12-24 | 삼성전자주식회사 | Robot cleaner and method for controlling the same |
KR101931362B1 (en) | 2011-08-22 | 2018-12-24 | 삼성전자주식회사 | Robot cleaner and method for controlling the same |
US20140058634A1 (en) | 2012-08-24 | 2014-02-27 | Crown Equipment Limited | Method and apparatus for using unique landmarks to locate industrial vehicles at start-up |
US8836751B2 (en) | 2011-11-08 | 2014-09-16 | Intouch Technologies, Inc. | Tele-presence system with a user interface that displays different communication links |
US9071603B1 (en) | 2011-12-27 | 2015-06-30 | Sirqul, Inc. | Distributed display functionality on multiple mobile devices |
US10054933B2 (en) | 2012-03-27 | 2018-08-21 | Sirqul, Inc. | Controlling distributed device operations |
US9146560B2 (en) | 2012-03-30 | 2015-09-29 | Irobot Corporation | System and method for implementing force field deterrent for robot |
US8902278B2 (en) | 2012-04-11 | 2014-12-02 | Intouch Technologies, Inc. | Systems and methods for visualizing and managing telepresence devices in healthcare networks |
US9251313B2 (en) | 2012-04-11 | 2016-02-02 | Intouch Technologies, Inc. | Systems and methods for visualizing and managing telepresence devices in healthcare networks |
US8965623B2 (en) * | 2012-05-11 | 2015-02-24 | International Business Machines Corporation | Automated cleaning in a sensor network |
US9361021B2 (en) | 2012-05-22 | 2016-06-07 | Irobot Corporation | Graphical user interfaces including touchpad driving interfaces for telemedicine devices |
WO2013176760A1 (en) | 2012-05-22 | 2013-11-28 | Intouch Technologies, Inc. | Graphical user interfaces including touchpad driving interfaces for telemedicine devices |
US9086274B2 (en) | 2012-06-27 | 2015-07-21 | Pentair Water Pool And Spa, Inc. | Pool cleaner with laser range finder system and method |
US9970788B2 (en) * | 2012-08-20 | 2018-05-15 | Dmg Mori Seiki Co., Ltd. | Scale measuring device, method for generating position information, and device with multi-axis stage |
EP3104194B1 (en) | 2012-08-27 | 2019-02-06 | Aktiebolaget Electrolux | Robot positioning system |
TWM451103U (en) * | 2012-10-30 | 2013-04-21 | Agait Technology Corp | Walking device |
CN103070638A (en) * | 2012-12-29 | 2013-05-01 | 北京石油化工学院 | Cleaning device capable of automatically walking |
CN109965778B (en) | 2013-01-18 | 2022-08-16 | 艾罗伯特公司 | Environment management system including mobile robot and method of using the same |
US9375847B2 (en) | 2013-01-18 | 2016-06-28 | Irobot Corporation | Environmental management systems including mobile robots and methods using same |
US9233472B2 (en) | 2013-01-18 | 2016-01-12 | Irobot Corporation | Mobile robot providing environmental mapping for household environmental control |
KR101450537B1 (en) | 2013-03-05 | 2014-10-14 | 엘지전자 주식회사 | Robot cleaner |
CN104053229B (en) * | 2013-03-14 | 2018-09-28 | 南京中兴软件有限责任公司 | Mobile terminal, localization method and device |
CN105101855A (en) | 2013-04-15 | 2015-11-25 | 伊莱克斯公司 | Robotic vacuum cleaner with protruding sidebrush |
WO2014169943A1 (en) | 2013-04-15 | 2014-10-23 | Aktiebolaget Electrolux | Robotic vacuum cleaner |
CN203987872U (en) | 2013-06-03 | 2014-12-10 | 碧洁家庭护理有限公司 | Autonomous type floor cleaning equipment |
CN103405208A (en) * | 2013-08-28 | 2013-11-27 | 徐振 | Sole dust cleaner |
WO2015094553A1 (en) | 2013-12-18 | 2015-06-25 | Irobot Corporation | Autonomous mobile robot |
US10209080B2 (en) | 2013-12-19 | 2019-02-19 | Aktiebolaget Electrolux | Robotic cleaning device |
WO2015090404A1 (en) | 2013-12-19 | 2015-06-25 | Aktiebolaget Electrolux | Prioritizing cleaning areas |
CN105813526B (en) | 2013-12-19 | 2021-08-24 | 伊莱克斯公司 | Robot cleaning device and method for landmark recognition |
WO2015090398A1 (en) | 2013-12-19 | 2015-06-25 | Aktiebolaget Electrolux | Robotic vacuum cleaner with side brush moving in spiral pattern |
JP6494118B2 (en) | 2013-12-19 | 2019-04-03 | アクチエボラゲット エレクトロルックス | Control method of robot cleaner associated with detection of obstacle climbing, and robot cleaner, program, and computer product having the method |
WO2015090402A1 (en) | 2013-12-19 | 2015-06-25 | Aktiebolaget Electrolux | Robotic cleaning device with perimeter recording function |
CN105744872B (en) | 2013-12-19 | 2020-01-14 | 伊莱克斯公司 | Adaptive speed control of rotating side brushes |
WO2015090439A1 (en) | 2013-12-20 | 2015-06-25 | Aktiebolaget Electrolux | Dust container |
JP5902664B2 (en) * | 2013-12-25 | 2016-04-13 | ファナック株式会社 | Human cooperative industrial robot with protective member |
US9283674B2 (en) | 2014-01-07 | 2016-03-15 | Irobot Corporation | Remotely operating a mobile robot |
JP6453539B2 (en) * | 2014-01-09 | 2019-01-16 | 東芝ライフスタイル株式会社 | Self-propelled equipment |
WO2015153109A1 (en) | 2014-03-31 | 2015-10-08 | Irobot Corporation | Autonomous mobile robot |
US10800038B1 (en) * | 2014-05-13 | 2020-10-13 | Al Incorporated | System and method for confinement of a robotic device |
US9731334B2 (en) | 2014-07-01 | 2017-08-15 | Kirt Ervin | Microbot pigging system |
ES2681802T3 (en) | 2014-07-10 | 2018-09-17 | Aktiebolaget Electrolux | Method to detect a measurement error in a robotic cleaning device |
CN107003656B (en) * | 2014-08-08 | 2020-02-07 | 机器人视觉科技股份有限公司 | Sensor-based safety features for robotic devices |
EP3190939B1 (en) | 2014-09-08 | 2021-07-21 | Aktiebolaget Electrolux | Robotic vacuum cleaner |
CN106659344B (en) | 2014-09-08 | 2019-10-25 | 伊莱克斯公司 | Robotic vacuum cleaner |
US9535153B2 (en) | 2014-09-09 | 2017-01-03 | Honeywell International Inc. | System and method for improved location accuracy |
US9510505B2 (en) | 2014-10-10 | 2016-12-06 | Irobot Corporation | Autonomous robot localization |
US9516806B2 (en) | 2014-10-10 | 2016-12-13 | Irobot Corporation | Robotic lawn mowing boundary determination |
CN105629972B (en) * | 2014-11-07 | 2018-05-18 | 科沃斯机器人股份有限公司 | Guiding virtual wall system |
US9788698B2 (en) | 2014-12-10 | 2017-10-17 | Irobot Corporation | Debris evacuation for cleaning robots |
US10877484B2 (en) | 2014-12-10 | 2020-12-29 | Aktiebolaget Electrolux | Using laser sensor for floor type detection |
CN107072454A (en) | 2014-12-12 | 2017-08-18 | 伊莱克斯公司 | Side brush and robot cleaner |
US9420741B2 (en) | 2014-12-15 | 2016-08-23 | Irobot Corporation | Robot lawnmower mapping |
WO2016095965A2 (en) | 2014-12-16 | 2016-06-23 | Aktiebolaget Electrolux | Experience-based roadmap for a robotic cleaning device |
JP6532530B2 (en) | 2014-12-16 | 2019-06-19 | アクチエボラゲット エレクトロルックス | How to clean a robot vacuum cleaner |
US9538702B2 (en) | 2014-12-22 | 2017-01-10 | Irobot Corporation | Robotic mowing of separated lawn areas |
TWI533101B (en) * | 2015-01-23 | 2016-05-11 | cheng-xiang Yan | System and Method of Restricting Robot Action |
USD775991S1 (en) | 2015-02-10 | 2017-01-10 | Irobot Corporation | Robot confinement beacon |
USD775990S1 (en) | 2015-02-10 | 2017-01-10 | Irobot Corporation | Robot confinement beacon |
US9757004B2 (en) | 2015-02-12 | 2017-09-12 | Irobot Corporation | Liquid management for floor-traversing robots |
US9993129B2 (en) | 2015-02-13 | 2018-06-12 | Irobot Corporation | Mobile floor-cleaning robot with floor-type detection |
US10545504B2 (en) | 2015-04-01 | 2020-01-28 | AI Incorporated | System and method for establishing virtual boundaries for robotic devices |
US9868211B2 (en) | 2015-04-09 | 2018-01-16 | Irobot Corporation | Restricting movement of a mobile robot |
US11099554B2 (en) | 2015-04-17 | 2021-08-24 | Aktiebolaget Electrolux | Robotic cleaning device and a method of controlling the robotic cleaning device |
US9746854B2 (en) | 2015-04-24 | 2017-08-29 | Autonomous Solutions, Inc. | System and method for controlling a vehicle |
JP6800952B2 (en) | 2015-04-24 | 2020-12-16 | アビドボッツ コーポレイション | Equipment and methods for semi-automatic surface cleaning |
US9505140B1 (en) | 2015-06-02 | 2016-11-29 | Irobot Corporation | Contact sensors for a mobile robot |
US9462920B1 (en) | 2015-06-25 | 2016-10-11 | Irobot Corporation | Evacuation station |
US9919425B2 (en) | 2015-07-01 | 2018-03-20 | Irobot Corporation | Robot navigational sensor system |
US11115798B2 (en) | 2015-07-23 | 2021-09-07 | Irobot Corporation | Pairing a beacon with a mobile robot |
US10034421B2 (en) | 2015-07-24 | 2018-07-31 | Irobot Corporation | Controlling robotic lawnmowers |
EP3344104B1 (en) | 2015-09-03 | 2020-12-30 | Aktiebolaget Electrolux | System of robotic cleaning devices |
US10021830B2 (en) | 2016-02-02 | 2018-07-17 | Irobot Corporation | Blade assembly for a grass cutting mobile robot |
US10459063B2 (en) | 2016-02-16 | 2019-10-29 | Irobot Corporation | Ranging and angle of arrival antenna system for a mobile robot |
JP7035300B2 (en) | 2016-03-15 | 2022-03-15 | アクチエボラゲット エレクトロルックス | Robot Cleaning Devices, Methods for Performing Escarpment Detection in Robot Cleaning Devices, Computer Programs, and Computer Program Products |
WO2017194102A1 (en) | 2016-05-11 | 2017-11-16 | Aktiebolaget Electrolux | Robotic cleaning device |
CN107402571A (en) * | 2016-05-20 | 2017-11-28 | 富泰华工业(深圳)有限公司 | Intensity of sound positioning guidance system and method |
US11263545B2 (en) | 2016-06-30 | 2022-03-01 | Microsoft Technology Licensing, Llc | Control of cyber-physical systems under uncertainty |
DE102016211842B4 (en) * | 2016-06-30 | 2024-04-25 | Robert Bosch Gmbh | Soil tillage equipment |
CN106200671A (en) * | 2016-07-19 | 2016-12-07 | 成都通甲优博科技有限责任公司 | A kind of comprehensive clean robot device and control method thereof |
EP3957447B1 (en) | 2016-09-14 | 2024-06-26 | iRobot Corporation | Systems and methods for configurable operation of a robot based on area classification |
US10632393B2 (en) * | 2017-01-13 | 2020-04-28 | Petronics Inc. | Mechanized tail for mobile devices |
JP2020515205A (en) | 2017-03-17 | 2020-05-21 | エスシーアールアールディー, インク.Scrrd, Inc. | Cross reference for wireless device detection, tracking and authentication platforms and technology applications |
US10341814B2 (en) * | 2017-03-17 | 2019-07-02 | SCRRD, Inc. | Wireless device detection, tracking, and authentication platform and techniques |
US10085118B1 (en) | 2017-03-17 | 2018-09-25 | SCRRD, Inc. | Wireless device detection, tracking, and authentication platform and techniques |
US11862302B2 (en) | 2017-04-24 | 2024-01-02 | Teladoc Health, Inc. | Automated transcription and documentation of tele-health encounters |
US10664502B2 (en) | 2017-05-05 | 2020-05-26 | Irobot Corporation | Methods, systems, and devices for mapping wireless communication signals for mobile robot guidance |
CN106974593A (en) * | 2017-05-12 | 2017-07-25 | 深圳星鸿云科技有限公司 | A kind of clean robot obstacle avoidance system and method |
JP7243967B2 (en) | 2017-06-02 | 2023-03-22 | アクチエボラゲット エレクトロルックス | Method for Detecting Level Differences on a Surface in Front of a Robotic Cleaning Device |
CN107102653B (en) * | 2017-06-07 | 2020-11-27 | 歌尔股份有限公司 | Device and method for controlling ground angle of mounting equipment of unmanned aerial vehicle |
US11470774B2 (en) | 2017-07-14 | 2022-10-18 | Irobot Corporation | Blade assembly for a grass cutting mobile robot |
US10483007B2 (en) | 2017-07-25 | 2019-11-19 | Intouch Technologies, Inc. | Modular telehealth cart with thermal imaging and touch screen user interface |
US11339580B2 (en) | 2017-08-22 | 2022-05-24 | Pentair Water Pool And Spa, Inc. | Algorithm for a pool cleaner |
US11636944B2 (en) | 2017-08-25 | 2023-04-25 | Teladoc Health, Inc. | Connectivity infrastructure for a telehealth platform |
US11921517B2 (en) | 2017-09-26 | 2024-03-05 | Aktiebolaget Electrolux | Controlling movement of a robotic cleaning device |
EP3702111B1 (en) * | 2017-10-25 | 2024-01-24 | LG Electronics Inc. | Artificial intelligence moving robot which learns obstacles, and control method therefor |
US10795377B2 (en) | 2018-01-03 | 2020-10-06 | AI Incorporated | Method for autonomously controlling speed of components and functions of a robot |
US11556125B2 (en) | 2018-01-05 | 2023-01-17 | Irobot Corporation | Mapping, controlling, and displaying networked devices with a mobile cleaning robot |
US10877156B2 (en) * | 2018-03-23 | 2020-12-29 | Veoneer Us Inc. | Localization by light sensors |
US11687092B2 (en) | 2018-04-23 | 2023-06-27 | Sharkninja Operating Llc | Techniques for bounding cleaning operations of a robotic surface cleaning device within a region of interest |
US10617299B2 (en) | 2018-04-27 | 2020-04-14 | Intouch Technologies, Inc. | Telehealth cart that supports a removable tablet with seamless audio/video switching |
US10595696B2 (en) | 2018-05-01 | 2020-03-24 | Sharkninja Operating Llc | Docking station for robotic cleaner |
EP3823507A4 (en) | 2018-07-20 | 2022-06-08 | SharkNinja Operating LLC | Robotic cleaner debris removal docking station |
US11191407B2 (en) | 2019-01-31 | 2021-12-07 | Irobot Corporation | Cleaning of pet areas by autonomous cleaning robots |
CN109828581B (en) * | 2019-02-27 | 2024-07-26 | 同济大学 | Multi-mode intelligent electric sweeper with automatic driving system |
CN109901590B (en) | 2019-03-30 | 2020-06-05 | 珠海市一微半导体有限公司 | Recharging control method of desktop robot |
CN110279350A (en) * | 2019-06-20 | 2019-09-27 | 深圳市银星智能科技股份有限公司 | From mobile device moving method and from mobile device |
ES2887526A1 (en) * | 2020-06-22 | 2021-12-22 | Cecotec Res And Development | SELF-MOVING DISINFECTION APPARATUS (Machine-translation by Google Translate, not legally binding) |
US20230255420A1 (en) | 2022-02-16 | 2023-08-17 | Irobot Corporation | Maintenance alerts for autonomous cleaning robots |
CN114603556A (en) * | 2022-03-07 | 2022-06-10 | 深圳创动科技有限公司 | Robot control method, control system, electronic device and readable storage device |
CN114415588B (en) * | 2022-03-25 | 2022-07-08 | 山东硅步机器人技术有限公司 | Adjustable monitoring control system |
Family Cites Families (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4119900A (en) | 1973-12-21 | 1978-10-10 | Ito Patent-Ag | Method and system for the automatic orientation and control of a robot |
US4679152A (en) * | 1985-02-20 | 1987-07-07 | Heath Company | Navigation system and method for a mobile robot |
NO864109L (en) | 1985-10-17 | 1987-04-21 | Knepper Hans Reinhard | PROCEDURE FOR AUTOMATIC LINING OF AUTOMATIC FLOOR CLEANING MACHINES AND FLOOR CLEANING MACHINE FOR PERFORMING THE PROCEDURE. |
US5109566A (en) | 1990-06-28 | 1992-05-05 | Matsushita Electric Industrial Co., Ltd. | Self-running cleaning apparatus |
JPH04227507A (en) | 1990-07-02 | 1992-08-17 | Nec Corp | Method for forming and keeping map for moving robot |
KR930000081B1 (en) | 1990-12-07 | 1993-01-08 | 주식회사 금성사 | Cleansing method of electric vacuum cleaner |
US5404461A (en) | 1991-03-29 | 1995-04-04 | International Business Machines Corp. | Broadcast/switching apparatus for executing broadcast/multi-cast transfers over unbuffered asynchronous switching networks |
DE4217093C1 (en) | 1992-05-22 | 1993-07-01 | Siemens Ag, 8000 Muenchen, De | |
US5548511A (en) | 1992-10-29 | 1996-08-20 | White Consolidated Industries, Inc. | Method for controlling self-running cleaning apparatus |
DE4408329C2 (en) * | 1994-03-11 | 1996-04-18 | Siemens Ag | Method for building up a cellular structured environment map of a self-moving mobile unit, which is oriented with the help of sensors based on wave reflection |
JPH0822322A (en) | 1994-07-07 | 1996-01-23 | Johnson Kk | Method and device for controlling floor surface cleaning car |
IL113913A (en) | 1995-05-30 | 2000-02-29 | Friendly Machines Ltd | Navigation method and system |
SE506907C2 (en) * | 1996-04-30 | 1998-03-02 | Electrolux Ab | Self-orientating device system and device |
AUPP299498A0 (en) * | 1998-04-15 | 1998-05-07 | Commonwealth Scientific And Industrial Research Organisation | Method of tracking and sensing position of objects |
US6282526B1 (en) * | 1999-01-20 | 2001-08-28 | The United States Of America As Represented By The Secretary Of The Navy | Fuzzy logic based system and method for information processing with uncertain input data |
US7155308B2 (en) * | 2000-01-24 | 2006-12-26 | Irobot Corporation | Robot obstacle detection system |
US7571511B2 (en) * | 2002-01-03 | 2009-08-11 | Irobot Corporation | Autonomous floor-cleaning robot |
US6883201B2 (en) * | 2002-01-03 | 2005-04-26 | Irobot Corporation | Autonomous floor-cleaning robot |
US6690134B1 (en) * | 2001-01-24 | 2004-02-10 | Irobot Corporation | Method and system for robot localization and confinement |
EP1395888B1 (en) * | 2001-06-12 | 2011-05-18 | iRobot Corporation | Method and system for multi-mode coverage for an autonomous robot |
US6658354B2 (en) * | 2002-03-15 | 2003-12-02 | American Gnc Corporation | Interruption free navigator |
ES2674568T3 (en) * | 2002-09-13 | 2018-07-02 | Irobot Corporation | Navigation control system for a robotic device |
-
2003
- 2003-09-12 ES ES16155717.8T patent/ES2674568T3/en not_active Expired - Lifetime
- 2003-09-12 EP EP16155717.8A patent/EP3043544B1/en not_active Expired - Lifetime
- 2003-09-12 EP EP03752285.1A patent/EP1547361B1/en not_active Expired - Lifetime
- 2003-09-12 US US10/661,835 patent/US7024278B2/en not_active Expired - Lifetime
- 2003-09-12 EP EP18164153.1A patent/EP3361716B1/en not_active Expired - Lifetime
- 2003-09-12 AU AU2003270581A patent/AU2003270581A1/en not_active Abandoned
- 2003-09-12 WO PCT/US2003/028624 patent/WO2004025947A2/en not_active Application Discontinuation
-
2006
- 2006-01-27 US US11/341,111 patent/US7188000B2/en not_active Expired - Lifetime
Also Published As
Publication number | Publication date |
---|---|
EP3361716A1 (en) | 2018-08-15 |
AU2003270581A8 (en) | 2004-04-30 |
EP1547361A4 (en) | 2012-05-30 |
EP1547361A2 (en) | 2005-06-29 |
EP3043544A1 (en) | 2016-07-13 |
EP3043544B1 (en) | 2018-03-28 |
US20040111184A1 (en) | 2004-06-10 |
WO2004025947A2 (en) | 2004-03-25 |
WO2004025947A9 (en) | 2005-04-14 |
US7188000B2 (en) | 2007-03-06 |
US7024278B2 (en) | 2006-04-04 |
ES2674568T3 (en) | 2018-07-02 |
AU2003270581A1 (en) | 2004-04-30 |
WO2004025947A3 (en) | 2004-05-21 |
US20060136096A1 (en) | 2006-06-22 |
EP1547361B1 (en) | 2016-04-06 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP3361716B1 (en) | Robotic cleaning device | |
US8718821B2 (en) | Navigational control system for a robotic device | |
US10813517B2 (en) | Navigational control system for a robotic device | |
US8386081B2 (en) | Navigational control system for a robotic device | |
EP2330473B2 (en) | Mobile robot | |
US7663333B2 (en) | Method and system for multi-mode coverage for an autonomous robot | |
US8396592B2 (en) | Method and system for multi-mode coverage for an autonomous robot |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE |
|
17P | Request for examination filed |
Effective date: 20180327 |
|
AC | Divisional application: reference to earlier application |
Ref document number: 1547361 Country of ref document: EP Kind code of ref document: P Ref document number: 3043544 Country of ref document: EP Kind code of ref document: P |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LI LU MC NL PT RO SE SI SK TR |
|
GRAP | Despatch of communication of intention to grant a patent |
Free format text: ORIGINAL CODE: EPIDOSNIGR1 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: GRANT OF PATENT IS INTENDED |
|
INTG | Intention to grant announced |
Effective date: 20201028 |
|
GRAS | Grant fee paid |
Free format text: ORIGINAL CODE: EPIDOSNIGR3 |
|
GRAA | (expected) grant |
Free format text: ORIGINAL CODE: 0009210 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE PATENT HAS BEEN GRANTED |
|
AC | Divisional application: reference to earlier application |
Ref document number: 3043544 Country of ref document: EP Kind code of ref document: P Ref document number: 1547361 Country of ref document: EP Kind code of ref document: P |
|
AK | Designated contracting states |
Kind code of ref document: B1 Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LI LU MC NL PT RO SE SI SK TR |
|
REG | Reference to a national code |
Ref country code: GB Ref legal event code: FG4D |
|
REG | Reference to a national code |
Ref country code: AT Ref legal event code: REF Ref document number: 1381189 Country of ref document: AT Kind code of ref document: T Effective date: 20210415 Ref country code: CH Ref legal event code: EP |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R096 Ref document number: 60352636 Country of ref document: DE |
|
REG | Reference to a national code |
Ref country code: IE Ref legal event code: FG4D |
|
REG | Reference to a national code |
Ref country code: NL Ref legal event code: MP Effective date: 20210407 Ref country code: AT Ref legal event code: MK05 Ref document number: 1381189 Country of ref document: AT Kind code of ref document: T Effective date: 20210407 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: AT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20210407 Ref country code: BG Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20210707 Ref country code: NL Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20210407 Ref country code: FI Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20210407 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: FR Payment date: 20210819 Year of fee payment: 19 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: SE Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20210407 Ref country code: PT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20210809 Ref country code: GR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20210708 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: DE Payment date: 20210827 Year of fee payment: 19 Ref country code: GB Payment date: 20210818 Year of fee payment: 19 |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R097 Ref document number: 60352636 Country of ref document: DE |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: SK Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20210407 Ref country code: EE Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20210407 Ref country code: ES Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20210407 Ref country code: RO Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20210407 Ref country code: DK Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20210407 Ref country code: CZ Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20210407 |
|
PLBE | No opposition filed within time limit |
Free format text: ORIGINAL CODE: 0009261 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT |
|
26N | No opposition filed |
Effective date: 20220110 |
|
REG | Reference to a national code |
Ref country code: CH Ref legal event code: PL |
|
REG | Reference to a national code |
Ref country code: BE Ref legal event code: MM Effective date: 20210930 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: MC Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20210407 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: LU Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20210912 Ref country code: IT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20210407 Ref country code: IE Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20210912 Ref country code: BE Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20210930 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: LI Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20210930 Ref country code: CH Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20210930 |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R119 Ref document number: 60352636 Country of ref document: DE |
|
GBPC | Gb: european patent ceased through non-payment of renewal fee |
Effective date: 20220912 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: CY Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20210407 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: HU Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT; INVALID AB INITIO Effective date: 20030912 Ref country code: FR Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20220930 Ref country code: DE Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20230401 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: GB Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20220912 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: TR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20210407 |