GB2568751A - Terrain analysis apparatus and method - Google Patents
Terrain analysis apparatus and method Download PDFInfo
- Publication number
- GB2568751A GB2568751A GB1719707.0A GB201719707A GB2568751A GB 2568751 A GB2568751 A GB 2568751A GB 201719707 A GB201719707 A GB 201719707A GB 2568751 A GB2568751 A GB 2568751A
- Authority
- GB
- United Kingdom
- Prior art keywords
- parking
- vehicle
- beacon
- reference pattern
- predetermined reference
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
- 238000004458 analytical method Methods 0.000 title claims abstract description 42
- 238000000034 method Methods 0.000 title claims description 47
- 230000003746 surface roughness Effects 0.000 claims description 10
- 230000008569 process Effects 0.000 claims description 9
- 230000006835 compression Effects 0.000 claims description 4
- 238000007906 compression Methods 0.000 claims description 4
- 238000012545 processing Methods 0.000 description 46
- 230000003287 optical effect Effects 0.000 description 23
- 238000001514 detection method Methods 0.000 description 20
- 230000003190 augmentative effect Effects 0.000 description 15
- 238000010586 diagram Methods 0.000 description 14
- 238000004891 communication Methods 0.000 description 13
- 230000000875 corresponding effect Effects 0.000 description 12
- 230000008859 change Effects 0.000 description 11
- 238000003032 molecular docking Methods 0.000 description 10
- 230000006870 function Effects 0.000 description 9
- 239000003550 marker Substances 0.000 description 9
- 238000000926 separation method Methods 0.000 description 9
- 238000003384 imaging method Methods 0.000 description 8
- 238000011217 control strategy Methods 0.000 description 7
- 230000008878 coupling Effects 0.000 description 7
- 238000010168 coupling process Methods 0.000 description 7
- 238000005859 coupling reaction Methods 0.000 description 7
- 230000033001 locomotion Effects 0.000 description 7
- 230000007246 mechanism Effects 0.000 description 7
- 230000002441 reversible effect Effects 0.000 description 7
- 230000001276 controlling effect Effects 0.000 description 6
- 238000012937 correction Methods 0.000 description 6
- 230000005540 biological transmission Effects 0.000 description 5
- 230000000007 visual effect Effects 0.000 description 4
- 244000025254 Cannabis sativa Species 0.000 description 3
- 230000004913 activation Effects 0.000 description 3
- 230000001447 compensatory effect Effects 0.000 description 3
- 230000001419 dependent effect Effects 0.000 description 3
- 238000011161 development Methods 0.000 description 3
- 230000001788 irregular Effects 0.000 description 3
- 239000011435 rock Substances 0.000 description 3
- 230000003213 activating effect Effects 0.000 description 2
- 230000003466 anti-cipated effect Effects 0.000 description 2
- 230000001413 cellular effect Effects 0.000 description 2
- 238000002485 combustion reaction Methods 0.000 description 2
- 238000004146 energy storage Methods 0.000 description 2
- 239000000463 material Substances 0.000 description 2
- 238000005259 measurement Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000001429 visible spectrum Methods 0.000 description 2
- 238000013459 approach Methods 0.000 description 1
- 230000001174 ascending effect Effects 0.000 description 1
- 239000010426 asphalt Substances 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 230000001351 cycling effect Effects 0.000 description 1
- 230000004069 differentiation Effects 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 238000005286 illumination Methods 0.000 description 1
- 238000010191 image analysis Methods 0.000 description 1
- 238000003780 insertion Methods 0.000 description 1
- 230000037431 insertion Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 244000144972 livestock Species 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 239000004576 sand Substances 0.000 description 1
- 239000000758 substrate Substances 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
- 230000001052 transient effect Effects 0.000 description 1
Classifications
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R1/00—Optical viewing arrangements; Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/20—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/22—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle
- B60R1/23—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle with a predetermined field of view
- B60R1/26—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle with a predetermined field of view to the rear of the vehicle
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/521—Depth or shape recovery from laser ranging, e.g. using interferometry; from the projection of structured light
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B62—LAND VEHICLES FOR TRAVELLING OTHERWISE THAN ON RAILS
- B62D—MOTOR VEHICLES; TRAILERS
- B62D13/00—Steering specially adapted for trailers
- B62D13/06—Steering specially adapted for trailers for backing a normally drawn trailer
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B62—LAND VEHICLES FOR TRAVELLING OTHERWISE THAN ON RAILS
- B62D—MOTOR VEHICLES; TRAILERS
- B62D15/00—Steering not otherwise provided for
- B62D15/02—Steering position indicators ; Steering position determination; Steering aids
- B62D15/027—Parking aids, e.g. instruction means
- B62D15/0285—Parking performed automatically
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01B—MEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
- G01B11/00—Measuring arrangements characterised by the use of optical techniques
- G01B11/24—Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures
- G01B11/25—Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures by projecting a pattern, e.g. one or more lines, moiré fringes on the object
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/003—Bistatic lidar systems; Multistatic lidar systems
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/88—Lidar systems specially adapted for specific applications
- G01S17/89—Lidar systems specially adapted for specific applications for mapping or imaging
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/88—Lidar systems specially adapted for specific applications
- G01S17/93—Lidar systems specially adapted for specific applications for anti-collision purposes
- G01S17/931—Lidar systems specially adapted for specific applications for anti-collision purposes of land vehicles
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R1/00—Optical viewing arrangements; Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/20—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/22—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle
- B60R1/23—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle with a predetermined field of view
- B60R1/27—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle with a predetermined field of view providing all-round vision, e.g. using omnidirectional cameras
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/30—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of image processing
- B60R2300/301—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of image processing combining image information with other obstacle sensor information, e.g. using RADAR/LIDAR/SONAR sensors for estimating risk of collision
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/30—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of image processing
- B60R2300/302—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of image processing combining image information with GPS information or vehicle data, e.g. vehicle speed, gyro, steering angle data
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/30—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of image processing
- B60R2300/304—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of image processing using merged images, e.g. merging camera image with stored images
- B60R2300/305—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of image processing using merged images, e.g. merging camera image with stored images merging camera image with lines or icons
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/40—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the details of the power supply or the coupling to vehicle components
- B60R2300/406—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the details of the power supply or the coupling to vehicle components using wireless transmission
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/80—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement
- B60R2300/8066—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement for monitoring rearward traffic
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/80—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement
- B60R2300/8086—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement for vehicle path indication
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30248—Vehicle exterior or interior
- G06T2207/30252—Vehicle exterior; Vicinity of vehicle
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Mechanical Engineering (AREA)
- Computer Networks & Wireless Communication (AREA)
- Electromagnetism (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Optics & Photonics (AREA)
- Combustion & Propulsion (AREA)
- Transportation (AREA)
- Chemical & Material Sciences (AREA)
- Theoretical Computer Science (AREA)
- Traffic Control Systems (AREA)
Abstract
A terrain feature analysis apparatus comprising a terrain feature analysis system with a processor configured to receive image data from one or more image sensors. The image data includes an image of a section of ground onto which a predetermined reference pattern 355 is projected by a portable beacon or projector 305. The image data is processed to identify an incident pattern 355 generated by the projection of the predetermined reference pattern 355 onto the section of ground. Distortions in the incident pattern 355 caused by projecting the predetermined reference pattern 355 onto the section of ground are identified and used to identify one or more terrain features in said section of ground. The system may be used on vehicles and used to assist parking by projecting a parking location indicator or vehicle trajectory indicator.
Description
TERRAIN ANALYSIS APPARATUS AND METHOD
TECHNICAL FIELD
The present disclosure relates to terrain analysis apparatus and method. More particularly, but not exclusively, the present disclosure relates to a terrain modelling apparatus; and a method of analysing terrain features.
BACKGROUND
It is known to provide vehicle parking systems for parking a trailer coupled to a vehicle. A target route may be generated to guide the trailer to a selected parking location. A vehicle control signal is output to control the vehicle, thereby guiding the trailer to the parking location. A potential shortcoming of known systems are the challenges associated in identifying a suitable parking location as the user may have limited information to assess factors such as surface condition or gradient.
At least in certain embodiments, the present invention seeks to ameliorate or overcome at least some of the aforementioned problems.
SUMMARY OF THE INVENTION
Aspects of the present invention relate to a terrain feature analysis system; a vehicle; a beacon; a method of analysing terrain features; and a non-transitory computer-readable medium as claimed in the appended claims.
According to a further aspect of the present invention there is provided a terrain feature analysis system for use with a vehicle, the terrain feature analysis system comprising a processor configured to:
receive image data from one or more image sensors, the image data comprising an image of a section of ground onto which a predetermined reference pattern is projected;
process the image data to identify an incident pattern generated by the projection of the predetermined reference pattern onto the section of ground;
identify distortions in the incident pattern caused by projecting the predetermined reference pattern onto the section of ground; and identify one or more terrain features in said section of ground in dependence on the identified distortions. The predetermined reference pattern is projected onto the section of ground. The predetermined reference pattern is an optical projection having a known profile. Image data comprising an image of that section of ground is collated and the incident pattern identified. The profile of the incident pattern is determined and any distortions identified. The distortions may correspond to differences between the predetermined reference pattern and the incident pattern identified in the image data. Each distortion may be indicative of a terrain feature which has altered the shape or profile of the predetermined reference pattern. By analysing the distortions, the processor may identify corresponding terrain features. When incorporated on a vehicle, the terrain feature analysis system allows the contours of the ground to be analysed, for example to identify changes in incline or gradient. The processor may be configured to control one or more vehicle systems in dependence on the identified terrain features, for example to control a torque output by a torque generating machine.
The processor may comprise an image processing module to analyse the image data and to identify the incident image. For example, an image processor may identify the incident pattern by applying a contrast filter or a colour filter to isolate the incident pattern.
The identification of distortions in said incident pattern may comprise comparing the incident pattern derived from the image data with the predetermined reference pattern. The predetermined reference pattern may comprise a projection of a predetermined pattern onto a surface having a known configuration. For example, the predetermined pattern may be projected onto a substantially flat surface which is at least substantially horizontal. Any distortions in the incident image may represent terrain features which differ from the known surface configuration.
The predetermined reference pattern could be determined during a calibration process. Alternatively, the predetermined reference pattern may be predefined.
The processor may be configured to identify one or more of the following distortions: compression of the predetermined reference pattern, extension of the predetermined reference pattern; skewing of the predetermined reference pattern; discontinuities in the predetermined reference pattern.
The predetermined reference pattern may be in the form of a terrain reference pattern. The predetermined reference pattern may comprise a plurality of regular or irregular geometric shapes, such as lines, curves, polygons (triangles, squares, pentagons, hexagons, etc.), circles or ellipses. The geometric shapes may have predetermined dimensions. The geometric shapes may be arranged in a predetermined configuration, for example at a known distance apart from each other. The geometric shapes may, for example, be arranged in an array having a predetermined spacing therebetween. The predetermined reference pattern may comprise a geometric pattern. The geometric pattern may comprise a mesh. The mesh may comprise a plurality of polygons, for example triangles, rectangles or hexagons. The mesh may comprise the same or different polygons. The dimensions of each polygon may be predefined. The polygons making up the mesh may have the same dimensions or different dimensions.
The one or more terrain features may comprise one or more of the following: an incline; a depression; a mound; surface roughness; and an object.
The predetermined reference pattern may be encoded. The encoding of the predetermined reference pattern may, for example, comprise changing the intensity of light emitted to form the predetermined reference pattern. The encoding of the predetermined reference pattern may facilitate identification of the incident pattern. The identification of the incident pattern may comprise identifying encoded image data.
The predetermined reference pattern may be generated by a projection system disposed in a beacon. The image may be captured by one or more image sensors disposed in a vehicle.
According to a further aspect of the present invention there is provided a comprising a terrain feature analysis system as described herein.
According to a further aspect of the present invention there is provided a beacon comprising a projection system for projecting a predetermined reference pattern onto a section of ground.
The beacon may comprise one or more light sources for emitting light to generate the predetermined reference pattern. At least one of the one or more light sources may be configured to encode the predetermined reference pattern.
The one or more light sources may comprise a laser.
The beacon may comprise a controllable mirror for reflecting light emitted from said one or more light sources to generate said predetermined reference pattern.
The beacon may be portable. The beacon may comprise a power supply, such as a battery.
The projection system may be configured to project a parking location indicator and/or a vehicle trajectory indicator.
According to a further aspect of the present invention there is provided a method of analysing terrain features for use with a vehicle, the method comprising:
projecting a predetermined reference pattern onto a section of ground;
capturing an image of the section of ground onto which the predetermined reference pattern is projected;
analysing the image to identify an incident pattern generated by the projection of the predetermined reference pattern onto the section of ground;
identifying distortions in the incident pattern caused by projecting the predetermined reference pattern onto the section of ground; and identifying one or more terrain features in said section of ground in dependence on the identified distortions.
The method may comprise identifying distortions in said incident pattern by comparing the incident pattern derived from the image data with the predetermined reference pattern.
The method may comprise identifying one or more of the following distortions: compression of the predetermined reference pattern, extension of the predetermined reference pattern; skewing of the predetermined reference pattern; discontinuities in the predetermined reference pattern.
The predetermined reference pattern may be in the form of a terrain reference pattern. The predetermined reference pattern may comprise a plurality of regular or irregular geometric shapes, such as lines, curves, polygons (triangles, squares, pentagons, hexagons, etc.), circles or ellipses. The geometric shapes may have predetermined dimensions. The geometric shapes may be arranged in a predetermined configuration, for example at a known distance apart from each other. The geometric shapes may, for example, be arranged in an array having a predetermined spacing therebetween. The predetermined reference pattern may comprise a geometric pattern. The geometric pattern may comprise a mesh. The mesh may comprise a plurality of polygons, for example triangles, rectangles or hexagons. The mesh may comprise the same or different polygons. The dimensions of each polygon may be predefined. The polygons making up the mesh may have the same dimensions or different dimensions.
The one or more terrain features may comprise one or more of the following: an incline; a depression; a mound; surface roughness; and an object.
The predetermined reference pattern may be generated by a projection system disposed in a beacon.
The image may be captured by one or more image sensors disposed in a vehicle.
The one or more terrain features may comprise one or more of the following: an incline; a depression; a mound; surface roughness; and obstacles.
According to a further aspect of the present invention there is provided a non-transitory computer-readable medium having a set of instructions stored therein which, when executed, cause a processor to perform the method described herein.
The one or more terrain features may comprise one or more of the following: an incline; a depression; a mound; surface roughness; and obstacles.
According to a further aspect of the present invention there is provided a non-transitory computer-readable medium having a set of instructions stored therein which, when executed, cause a processor to perform the method described herein.
It is to be understood that references herein to terrain features include one or more of the following set: incline; gradient (ascending or descending); dips; depressions; holes; drops; ruts; steps; rocks; surface roughness; (impassable) obstructions; and surface composition. In addition to identification of the terrain features, the method and apparatus described herein may grade the terrain features. The terrain features may be graded in dependence on one or more of the following set: size, severity and prevalence. The techniques described herein may also reference the type of terrain to be traversed. It will be understood that the term 'type of terrain' refers to the material comprised by the terrain, such as asphalt, grass, gravel, snow, mud, rock and/or sand. The type of terrain may optionally provide an indication of a coefficient of friction of the surface. The type of terrain may be determined using one or more sensor or may be specified by a user.
Any control unit or controller described herein may suitably comprise a computational device having one or more electronic processors. The system may comprise a single control unit or electronic controller or alternatively different functions of the controller may be embodied in, or hosted in, different control units or controllers. As used herein the term “controller” or “control unit” will be understood to include both a single control unit or controller and a plurality of control units or controllers collectively operating to provide any stated control functionality. To configure a controller or control unit, a suitable set of instructions may be provided which, when executed, cause said control unit or computational device to implement the control techniques specified herein. The set of instructions may suitably be embedded in said one or more electronic processors. Alternatively, the set of instructions may be provided as software saved on one or more memory associated with said controller to be executed on said computational device. The control unit or controller may be implemented in software run on one or more processors. One or more other control unit or controller may be implemented in software run on one or more processors, optionally the same one or more processors as the first controller. Other suitable arrangements may also be used.
Within the scope of this application it is expressly intended that the various aspects, embodiments, examples and alternatives set out in the preceding paragraphs, in the claims and/or in the following description and drawings, and in particular the individual features thereof, may be taken independently or in any combination. That is, all embodiments and/or features of any embodiment can be combined in any way and/or combination, unless such features are incompatible. The applicant reserves the right to change any originally filed claim or file any new claim accordingly, including the right to amend any originally filed claim to depend from and/or incorporate any feature of any other claim although not originally claimed in that manner.
BRIEF DESCRIPTION OF THE DRAWINGS
One or more embodiments of the present invention will now be described, by way of example only, with reference to the accompanying figures, in which:
Figure 1 shows a park assist system in accordance with an embodiment of the present invention;
Figure 2 shows a schematic representation of a vehicle incorporating a park assist controller for use in the park assist system shown in Figure 1;
Figure 3 shows a schematic representation of a parking beacon for use in the park assist system shown in Figure 1;
Figure 4A shows a block diagram representing operation of the parking assist system to project an encoded parking location indicator;
Figure 4B shows a block diagram representing operation of the parking assist system to generate a target route;
Figure 5 illustrates a target route generated by the park assist system for guiding a trailer to a target position;
Figure 6A shows a display image from a rear camera with a graphical overlay representing the target position of the trailer;
Figure 6B shows a birds-eye view image including features of the trailer to be projected along with the parking location indicator;
Figure 7 shows a block diagram representing operation of the parking assist system to generate a graphical representation of the target route;
Figure 8 illustrates the vehicle and trailer dimensions and associated nomenclature for determining the target route for the trailer;
Figure 9 shows a schematic representation of a vehicle incorporating a park assist controller in accordance with a further embodiment of the present invention;
Figure 10 illustrates a target route generated by the park assist system for guiding the host vehicle to a target position;
Figure 11 shows a schematic representation of a vehicle incorporating a park assist controller in accordance with a further embodiment of the present invention;
Figure 12 illustrates a target route generated by the park assist system for guiding a remote vehicle to the target position;
Figure 13 shows a block diagram representing operation of the parking assist system to transmit a target route to the remote vehicle;
Figure 14 shows a schematic representation of an alternative parking beacon for use in accordance with the present invention;
Figure 15 illustrates a target route generated by the park assist system utilising the parking beacon illustrated in Figure 12;
Figure 16A illustrates the operation of the parking beacon shown in Figure 12 to project a terrain reference pattern onto a flat surface;
Figure 16B illustrates the operation of the parking beacon shown in Figure 12 to project a terrain reference pattern onto an uneven surface;
Figure 17 shows a block diagram representing operation of the parking assist system to analyse a projected reference pattern to identify terrain features;
Figure 18 illustrates the operation of the parking beacon shown in Figure 12 to project a terrain reference pattern between the target position and the current position of the towed vehicle;
Figure 19 represents a torque profile required to propel the towed vehicle to the target position based on analysis of the terrain reference pattern;
Figure 20 shows a block diagram representing operation of the parking assist system to generate a torque control strategy;
Figure 21A illustrates the operation of the parking beacon shown in Figure 12 to project a vehicle visible warning indicator onto a flat surface;
Figure 21B illustrates the operation of the parking beacon shown in Figure 12 to project a vehicle visible warning indicator onto an uneven surface;
Figure 22A illustrates the operation of the parking beacon shown in Figure 12 to project path lines onto a flat surface;
Figure 22B illustrates the operation of the parking beacon shown in Figure 12 to project path lines onto an uneven surface;
Figure 23 shows a block diagram representing operation of the parking assist system to analyse a projected route indicator to identify terrain features;
Figure 24 shows a schematic representation of a parking beacon incorporating a parking assist controller in accordance with an embodiment of the present invention;
Figure 25 shows a schematic representation of an exemplary parking scenario utilising the parking beacon shown in Figure 24;
Figure 26 shows a schematic representation of a first augmented image generated by the parking beacon shown in Figure 25;
Figure 27 shows a schematic representation of a second augmented image generated by the parking beacon shown in Figure 25;
Figure 28 shows a schematic representation of a docking station for the parking beacon shown in Figure 25; and
Figure 29 shows a schematic representation of an exemplary parking scenario utilising a variant of the parking beacon shown in Figure 24.
DETAILED DESCRIPTION
A parking assist system 1 in accordance with an aspect of the present invention will now be described with reference to the accompanying Figures. The parking assist system 1 according to the present embodiment is operative to assist manoeuvring of a towing vehicle and a towed vehicle 3 coupled together by an articulated coupling. The parking assist system 1 has particular application in assisting in the performance of a reversing manoeuvre, for example to perform a reverse parking manoeuvre to park the towed vehicle (and optionally also the towing vehicle 2). The reverse parking manoeuvre may, for example, comprise parallel parking or tandem parking.
As described herein, the parking assist system 1 comprises a parking assist controller 4 and a parking beacon 5. The parking assist controller 4 in the present embodiment is installed in the towing vehicle 2. The towing vehicle 2 is illustrated herein as an automobile, such as a utility vehicle or a sports utility vehicle (SUV), but it will be understood that the parking assist system 1 described herein may be used in other vehicle types, such as a tractor. The towed vehicle 3 in the present embodiment comprises a trailer, such as a caravan. The parking beacon 5 in the present embodiment is a portable device which is adapted to be positioned externally of the towed vehicle 3 to indicate a target position for parking the towed vehicle 3.
A plan view of the towing vehicle 2 and the towed vehicle 3 is shown in Figures 1 and 2. The towing vehicle 2 has a first longitudinal axis X1 and the towed vehicle 3 has a second longitudinal axis X2. As shown in Figure 2, the towing vehicle 2 has two front wheels W1, W2 and two rear wheels W3, W4. As illustrated in Figure 2, the front wheels W1, W2 of the towing vehicle 2 are steerable in conventional manner to define a steering angle Θ. The steering angle Θ is the orientation of the front wheels W1, W2 relative to the longitudinal axis X1 of the towing vehicle. The towed vehicle 3 in the present embodiment has two trailer wheels TW1, TW2. The trailer wheels TW1, TW2 are mounted on a single axle and are not steerable. The parking assist controller 4 may be used in conjunction with a towed vehicle 3 having more than one axle, for example a twin-axle trailer, and/or more than two trailer wheels.
As shown in Figures 1 and 2, the towing vehicle 2 is provided with a vehicle imaging system comprising a central towing vehicle camera 8, a left towing vehicle camera 9 and a right towing vehicle camera 10. The central towing vehicle camera 8 is mounted in a central position at the rear of the towing vehicle 2, for example in a rear bumper. The left towing vehicle camera 9 and the right towing vehicle camera 10 are mounted in respective left and right wing mirrors 11, 12 of the towing vehicle 2. The towing vehicle cameras 8, 9, 10 are optical cameras arranged to face in a rearwards direction, i.e. towards a region disposed at the rear of the towing vehicle 2. The fields of view V1, V2, V3 of the central towing vehicle camera 8, the left towing vehicle camera 9 and the right towing vehicle camera 10 are illustrated by dashed lines in Figures 1 and 2. The towing vehicle 2 also comprises an obstruction detection system for detecting the presence of an obstruction in the rear three quarters position of the towing vehicle 2. The obstruction detection system in the present embodiment comprises left and right radar systems 13A, 13B. The left and right radar sensors 13A, 13B are operative to detect obstructions in respective left and right detection regions 14A, 14B.
As shown in Figure 1, the towed vehicle 3 is provided with a trailer imaging system comprising a rear trailer camera 15. The trailer imaging system may optionally comprise additional trailer cameras, for example one or more trailer cameras mounted on the sides of the towed vehicle 3. The rear trailer camera 15 is an optical camera arranged to provide a video image of an area behind the towed vehicle 3 which may otherwise be obscured from the driver’s field of vision. The rear trailer camera 15 has a field of view TV1 which is illustrated by dashed lines. The towed vehicle 3 optionally also comprises an obstruction detection system, for example parking sensors 16A, 16B, for detecting obstructions in one or more detection regions at the rear of the towed vehicle 3 and/or at the sides of the towed vehicle 3. The image data from the rear trailer camera 15 and/or obstruction data from the obstruction detection system is transmitted to the towing vehicle 2 over a wired connection or a wireless connection.
The coupling between the towing vehicle 2 and the towed vehicle 3 is an articulated coupling. In the present embodiment, a tow hitch 17 is mounted to the towing vehicle 2 for connection to a trailer coupling 19 mounted to the towed vehicle 3. The tow hitch 17 in the present embodiment comprises an upwardly projecting tow ball. The trailer coupling 19 is mounted to a hitch frame 21 disposed at the front of the towed vehicle 3. In the present embodiment, the hitch frame 21 is an A-frame having a front apex to which the trailer coupling 19 is mounted. The angular orientation of the first longitudinal axis X1 of the towing vehicle 2 and the second longitudinal axis X2 of the towed vehicle 3 is referred to herein as a hitch angle φ. The towing vehicle 2 comprises a hitch angle determining means 24 for measuring the hitch angle φ. The hitch angle determining means 24 may comprise, for example, a mechanical sensor or an optical sensor. In the present embodiment, the hitch angle determining means 24 comprises an optical system for determining an angular orientation of a target 25 mounted to of the towed vehicle 3. The target 25 is mounted on the towed vehicle 3 within the field of view V1 of the central towing vehicle camera 8. The target 25 is a visible image comprising three circles arranged in a triangular formation. It will be appreciated that the present invention can be implemented with other targets 25, for example comprising different symbols/images or non-visible targets.
The parking assist controller 4 is configured to control the towing vehicle 2 to assist manoeuvring of the towing vehicle 2 and the towed vehicle 3. The parking assist controller 4 is operative selectively to control the towing vehicle 2 to manoeuvre the combination of the towing vehicle 2 and the towed vehicle 3 to position the towed vehicle 3 (and optionally also the towing vehicle 2) in a predetermined target position Ptar. The target position Ptar defines a target location and/or a target orientation of the towed vehicle 3 for parking. As described herein, the parking assist controller 4 is operative to generate a parking assist signal S1 to control the towing vehicle 2 to guide the towed vehicle 3 to the target position Ptar. The parking assist signal S1 provides automated or semi-automated control of the towing vehicle 2 to park the towed vehicle 3. In particular, the parking assist signal S1 is output to an electronic power assisted steering (EPAS) module 18 to control the steering angle θ of the front wheels W1, W2 of the towing vehicle 2. The parking assist signal S1 may optionally control a torque request signal which is output to a torque generating machine, such as an internal combustion engine or an electric machine; and/or a brake torque request signal to a braking apparatus. Alternatively, the parking assist signal S1 may be output to a human machine interface (HMI), for example comprising a display screen or a head up display (HUD), to provide instructions or prompts to the driver of the towing vehicle 2.
In known arrangements, the target position Ptar may be specified using an HMI, for example by positioning a target parking position on a screen. A potential limitation of known arrangements is that it may prove difficult for a user to visualise the location of the towed vehicle 3 before completing the parking manoeuvre. The user may find it difficult to relate onscreen features on a display screen with the real-world counterparts. By way of example, if the towed vehicle 3 is a caravan, it may prove difficult for a user to align a charging point on the towed vehicle 3 with a mains electrical supply shown on a display screen. At least in certain embodiments, the parking assist controller 4 according to the present invention may overcome or reduce this problem. In particular, the parking assist controller 4 is configured to determine the target position Ptar with reference to a parking location indicator 26 generated by a parking beacon 5. As described herein, the parking location indicator 26 in the present embodiment is projected onto the ground by one or more optical systems provided in the parking beacon 5.
The parking beacon 5 is illustrated in Figure 3. The parking beacon 5 is a portable device which may be free standing, for example comprising a tripod structure; or may comprise a spike (not shown) for insertion into the ground. The parking beacon 5 comprises a beacon control unit 27 and at least a first projection system 28 for projecting the parking location indicator 26 onto the ground. The first projection system 28 in the illustrated arrangement comprises a light source 29; and optical guide means 30. The optical guide means 30 is configured to define a light path for the light emitted from the light source 29. The optical guide means 30 in the present embodiment is controllable to adjust the light path of the light emitted from the light source 29. The optical guide means 30 in the present embodiment comprises a mirror 30. Other forms of optical guide may be used, such as a reflective prism. The light source 29 is configured to emit at least one light path 31 of visible light for projecting the parking location indicator 26. In use, the at least one light path 31 may be projected onto the ground so that the parking location indicator 26 is visible. The beacon control unit 27 controls the energization of the light source 29 and the orientation of the mirror 30 to form the parking location indicator 26. By changing the orientation of the mirror 30, the position and/or orientation and/or profile of the parking location indicator 26 may be controlled dynamically. The at least one light path 31 provides a visual representation of a position and/or orientation and/or profile of the target position Ptar for the towed vehicle 3. The light source 29 is described herein as emitting a single light path 31 of light, but the parking beacon 5 may emit multiple light paths 31, for example to represent two or more sides of the target position Ptar. The first projection system 28 may optionally include optics, such as one or more lenses, to focus the light emitted from the light source 29.
The light source 29 comprises at least one laser diode. The mirror 30 is an optical mirror configured to reflect the light emitted from the light source 29. The mirror 30 can be oriented so as to direct the light path 31 downwardly (relative to a vertical axis of the parking beacon 5) so that is incident on the ground. The mirror 30 in the present embodiment is a scanning mirror and the orientation of the mirror 30 is adjustable dynamically to modulate the emitted light. The mirror 30 may, for example, comprise a micro-opto-electromechanical system (MOEMS). The mirror 30 pivots about at least one axis to modulate the emitted light, thereby forming the visible representation of the parking location indicator 26. In the present embodiment, the mirror 30 is pivotable about a first pivot axis X1 and a second pivot axis Z1, arranged perpendicular to each other, to enable the light path 31 to be controlled in first and second directions. The mirror 30 may optionally also be pivotable about a third pivot axis Y1, arranged perpendicular to the first pivot axis X1 and the second pivot axis Z1. The beacon control unit 27 is configured to control the orientation of the mirror 30 to trace the profile of the parking location indicator 26. By controlling the orientation of the mirror 30, the light path 31 may scan a visible pattern representing the parking location indicator 26. The beacon control unit 27 controls the mirror 30 to project the light path 31 onto the ground so that the parking location indicator 26 is visible. The beacon control unit 27 may selectively activate and deactivate the light source 29 to generate a continuous or interrupted trace. The parking location indicator 26 in the present embodiment comprises a rectangular frame corresponding to an outer perimeter of the target position Ptar. The beacon control unit 27 may optionally modulate the light source 29 to encode the light path 31. The modulation of the light source 29 may, for example, comprise selectively activating and deactivating in a predetermined pattern which is identifiable by the parking assist controller 4.
The beacon control unit 27 controls the energization of the light source 29 and the orientation of the mirror 30 to form the parking location indicator 26. By changing the orientation of the mirror 30, the parking location indicator 26 is projected onto the ground to provide a visual representation of the target position Ptar. The light path 31 projected by the parking beacon 5 may, for example, be controlled to define a side of the target position Ptar. The mirror 30 may be controlled to highlight one or more features, for example the extremities of the target position Ptar. As outlined above, the parking beacon 5 may generate more than one light path 31 of light. For example, the parking beacon 5 may emit two (2) light paths 31 of visible light suitable for representing adjacent or opposing boundaries of the target position Ptar. The light paths 31 could both be the same colour of light (same wavelength), or may be different colours of light (different wavelength) to differentiate between different features. Alternatively, or in addition, the parking assist system 1 described herein could comprise a plurality of parking beacons 5.
The operation of the beacon control unit 27 is illustrated with reference to a block diagram 70 shown in Figure 4A. The beacon control unit 27 determines the target position Ptar for parking the towed vehicle 3 (BLOCK A1). The beacon control unit 27 generates an image comprising the parking location indicator 26 (BLOCK A2). The image is encoded to facilitate identification (BLOCK A3). The beacon control unit 27 controls the first projection system 28 to project the encoded image (BLOCK A4).
The light source 29 and the mirror 30 may optionally be mounted to an adjustable mounting assembly 22. The mounting assembly 22 may enable manual or automatic adjustment in a in a vertical plane and/or rotation about a vertical axis. The adjustable mounting assembly 22 may optionally comprise a self-levelling apparatus, such as a gimbal. The self-levelling apparatus may ensure a reference level for the projection of the parking location indicator 26.
As outlined above, the parking location indicator 26 is visible and allows a user to position and/or orient the target position. In the arrangement described herein, the parking location indicator 26 comprises a single line which is traced by scanning the light path 31. This line defines a boundary of the parking location indicator 26 and is used to determine the target position Ptar for parking the towed vehicle 3. The parking beacon 5 may be modified to generate two (2) or more lines, for example arranged orthogonally to define a corner of the parking location indicator 26. Alternatively, the parking beacon 5 may generate a box or frame representing the boundaries of the parking location indicator 26. The beacon control unit 27 may be configured to match the size of the parking location indicator 26 to the actual size of the towed vehicle 3 and/or the towing vehicle 2 (for example by direct entry of the dimensions and/or transfer from the vehicle/trailer system). In a further modification, the beacon control unit 27 may control the mirror 30 to trace additional details, for example to indicate the location of features of the towed vehicle 3, such as a wheel(s), a drainage point(s), an electrical connection, a door, a hitch etc. Alternatively, or in addition, the parking location indicator 26 may include features representing the position of a lowered tailgate or a side door (for example, if the towed vehicle 3 is a horsebox). In a modified arrangement, the parking beacon 5 may project a plurality of spots (point illuminations) to represent the parking location indicator 26. The parking location indicator 26 may, for example, by represented by four (4) spots corresponding to the corners of the target position Ptar. By composing the parking location indicator 26 of one or more spots, the parking location indicator 26 may be detectable in a wider range of operating conditions, for example in bright sunlight. A less powerful laser diode may be used to form the parking location indicator 26.
The parking beacon 5 comprises a beacon human machine interface (HMI) 32 to enable the user to control operation of the parking beacon 5. The beacon HMI 32 may, for example, enable the user to specify a particular orientation and/or location of the parking location indicator 26 to be projected onto the ground. The beacon HMI 32 could be implemented by a software application operated on a separate computational device, such as a cellular telephone or personal computer, connected over a wired or wireless communication channel. It is envisaged that a short-range wireless communication channel would provide a suitable connection to the parking beacon 5. In the present embodiment, the beacon control unit 27 is operative to control the light source 29 to encode a signal into the parking location indicator 26. The encoded signal may be an identification signal to facilitate detection and/or identification of the parking location indicator 26. For example, to facilitate orientation, the projected lines may be pulsed to generate a code which differentiates between a longitudinal line and a lateral line. The pulse codes may be extended to convey other information such as trailer identification and/or trailer type and/or absolute dimension of the trailer.
Alternatively, or in addition, the encoded signal may comprise data relating to the dimensions and/or configuration of the towed vehicle 3. The parking beacon 5 may optionally comprise means for sensing the towed vehicle 3, such as a proximity sensor, to provide feedback to the parking assist controller 4. The parking beacon 5 may optionally comprise a camera for transmitting image data to the parking assist controller 4, for example to provide an alternate view of the region behind the towed vehicle 3.
The parking assist controller 4 provided in the towing vehicle 2 comprises an electronic control unit (ECU) 33 having an electronic processor 34 and a memory 35, as shown schematically in Figure 2. The electronic processor 34 is configured to execute a set of computational instructions stored on the memory 35. The processor 34 comprises image processing means for analysing image data generated by the towing vehicle cameras 8, 9, 10 and the trailer camera(s) 15.
The processor 34 comprises a first image processing module 36A for receiving the image data from the central towing vehicle camera 8. The first image processing module 36A is configured to analyse the image data to determine the position and/or orientation of the target 25 mounted to the towed vehicle 3. By determining the position and/or orientation of the target 25, the first image processing module 36A determines the hitch angle φ of the towed vehicle 3. The first image processing module 36A thereby performs the function of the hitch angle determining means 24 in the present embodiment. The first image processing module 36A thereby determines a current position Pcur of the towed vehicle 3.
The processor 34 comprises a second image processing module 36B for receiving image data from the left and right towing vehicle cameras 9, 10 and the rear trailer camera 15. The second image processing module 36B is configured to analyse the image data to identify the parking location indicator 26 generated by the parking beacon 5. In particular, the second image processing module 36B processes the image data to identify any elements in the image data which change in accordance with the encoded signal introduced by the beacon control unit 27. The second image processing module 36B may thereby identify the parking location indicator 26. The processor 34 determines the target position Ptar for parking the towed vehicle 3 in dependence on the parking location indicator 26. In the illustrated arrangement, the light path 31 generated by the parking beacon 5 indicates a side boundary of the target position Ptar and a longitudinal extent of the target position Ptar. In a variant, the light path 31 generated by the parking beacon 5 may represent a centreline of the target position Ptar. The ends of the parking location indicator 26 represent the front and rear extremities of the target position Ptar.
The processor 34 further comprises vehicle/trailer guidance means in the form of a guidance module 36C. The guidance module 36C is provided to assist with reversing the combination of the towing vehicle 2 and the towed vehicle 3. In particular, the guidance module 36C is configured to output a control signal for controlling the steering angle θ of the front wheels W1, W2 of the towing vehicle 2 to guide the towed vehicle 3 along a target route (or trajectory) R. The target route R is generated by the guidance module 36C to guide the towed vehicle 3 from a current position Pcur to the target position Ptar. The target route R is illustrated by a broken line in Figure 5. The guidance module 36C is configured to generate the target route R to guide the towed vehicle 3 from the current position Pcur to the identified target position Ptar. The target route R defines a trajectory for the towed vehicle 3 from the current position Pcur to the target position Ptar. The target route R can comprise rectilinear and/or curved sections. The target route R is arranged coincident with a midpoint of the towed vehicle 3 in the current position Pcur. The current position Pcur of the towed vehicle 3 is monitored compared to the originally calculated target route R. Small deviations are managed within the parking assist controller 4. Larger deviations can trigger a recalculation of the target route R. If the target position Ptar becomes unachievable from the current position PCur, the user is alerted and a corrective manoeuvre is suggested (for example travel forward a short distance).
The processor 34 comprises a third image processing module 36D configured to identify obstacles in the trajectory of the towing vehicle 2 and the towed vehicle 3. The third image processing module 36D may receive detection signals from the left and right radar systems 13A, 13B to identify the presence/absence of obstructions proximal to the towing vehicle 2 and the towed vehicle 3. In particular, the third image processing module 36D is configured to identify obstructions along the target route R of the towing vehicle 2 and/or the towed vehicle 3.
The operation of the processor 34 is illustrated with reference to a block diagram 71 shown in Figure 4B. The processor 34 receives the image data from the central towing vehicle camera 8 (BLOCK B1). The first image processing module 36A analyses the image data to determine a current position PCur of the towed vehicle 3 (BLOCK B2). The processor 34 receives the image data from the left and right towing vehicle cameras 9, 10 and the rear trailer camera 15 (BLOCK B3). The second image processing module 36B analyses the image data to identify the parking location indicator 26 (BLOCK B4). The processor 4 uses the parking location indicator 26 to determine the target position Ptar of the towed vehicle 3 (BLOCK B5). The guidance module 36C determines the target route R from the current position PCur to the target position Ptar (BLOCK B6). The third image processing module 36D identifies obstacles along the target route R (BLOCK B7). The third image processing module 36D may optionally receive additional image data from one or more cameras provided on the towing vehicle 2 and the towed vehicle 3 (BLOCK B8). If required, the guidance module 36C may modify the target route R to avoid any obstacles identified by the third image processing module 36D (BLOCK B9). It will be understood that any obstacles may be identified prior to generation of the target route R. The parking assist signal S1 comprising the modified target route R is output (BLOCK B10).
The parking assist controller 4 comprises a human machine interface (HMI) 37. The controller HMI 37 in the present embodiment is provided in the towing vehicle 2, but could be provided in the towed vehicle 3. The controller HMI 37 comprises a display screen 38 and is configured to display a target position indicator 39 representing the target position Ptar. In the present embodiment, the target position indicator 39 is superimposed onto an image 40 received from the central towing vehicle camera 8 and/or the rear trailer camera 15. As illustrated in Figure 6A, the target position indicator 39 is illustrated as a rectangle representing the footprint of the towing vehicle 2 and the towed vehicle 3 in combination. The target position indicator 39 is sized to provide a scale representation of the towing vehicle 2 and the towed vehicle 3 within the image 40. The dimensions of the towing vehicle 2 are defined in a data file accessible to the parking assist controller 4. To provide an accurate representation of the towed vehicle 3, the trailer dimensions can be specified by the user and stored in the data file. The controller HMI 37 may optionally comprise input means 41 which can be operated by the user to adjust the position of the target position indicator 39 within the image 40. The input means 41 could comprise a touch sensitive screen and/or a rotary dial, for example. If required, the user can adjust or refine the position of the target position Ptar determined by the second image processing module 36B. The controller HMI 37 could optionally also be configured to allow the user to adjust the target route R generated by the guidance module 36C.
The controller HMI 37 may optionally also enable the user to specify the position and/or size of one or more features of the towed vehicle 3. As shown in Figure 6B, the image 40 output by the controller HMI 37 may provide a graphical representation of the towed vehicle 3. The representation of the towed vehicle 3 may include one or more markers representing features of the towed vehicle. The image 40 may comprise one or more of the following set: a door marker 42, an electrical connector marker 43, an A-frame marker 44, a guide wheel (not shown) and a retractable (support) member (not shown). The image 40 may show the position and/or size of the features, for example in an open state and/or a closed state. The image 40 may, for example, comprise a birds-eye view of the towed vehicle 3. The birds-eye view may be a composite of multiple images captured by the cameras 8, 9, 10, provided on the towing vehicle 2 and/or the camera 15 provided on the towed vehicle 3. The controller HMI 37 may be configured to allow the user to define the various features of the towed vehicle 3. Alternatively, a model of the towed vehicle 3 may be pre-defined, for example stored in a database accessed by the parking assist controller 4. By including the features of the towed vehicle 3 in the image 40, the user may more readily select the target position Ptar. For example, the user may select the target position Ptar so as to position the electrical connector marker 43 proximal to a power point.
The operation of the controller HMI 37 is illustrated with reference to a block diagram 72 shown in Figure 7. The controller HMI 37 determines the target position Ptar (BLOCK C1). As described herein, the target position Ptar may be determined by analysing image data to identify the parking location indicator 26 projected by the first projection system 28 (BLOCK
C2). Alternatively, or in addition, the target position Ptar may be determined by user input (BLOCK C3). The controller HMI 37 receives the image data from the left and right towing vehicle cameras 9, 10 and the rear trailer camera 15 (BLOCK C4). The image data is displayed on the display screen 38 (BLOCK C5). The target position indicator 39 is overlaid onto the image (BLOCK C6). A user input is received, for example via user interaction with the touch screen, to adjust the position of the target position indicator 39 (BLOCK C7). In dependence on the received user inputs, the controller HMI 37 adjusts the location of the target position indicator 39 within the image (BLOCK C8). The user validates the location of the target position indicator 39 and, as described herein, the target route R is generated (BLOCK C9). The target route R may then be overlaid onto the image to provide a graphical representation (BLOCK C10). The target route R may optionally be modified, for example in dependence on user inputs (BLOCK C11).
The first projection system 28 may be configured to project the features of the towed vehicle 3 along with the parking location indicator 26. For example, the first projection system 28 may project the features onto the ground to facilitate selection of the target position Ptar for the towed vehicle 3.
In use, the user positions the parking beacon 5 externally of the towing vehicle 2 and the towed vehicle 3. The parking beacon 5 is typically placed on the ground. The first projection system 28 is activated to project the parking location indicator 26. The parking location indicator 26 is projected onto the ground to provide an at least substantially full-size representation (i.e. a 1:1 representation) of the target position (Ptar). The user configures the first projection system 28 such that parking location indicator 26 corresponds to a desired target position Ptar. The user may manually configure the first projection system 28 or may utilise the HMI 37 to set the location of the parking location indicator 26. The parking location indicator 26 is projected onto the ground and thereby provides a visible indication of the target position Ptar. In the present embodiment, the parking location indicator 26 represents a boundary of the target position Ptar.
The guidance module 36C determines a current position Pcur of the towed vehicle 3 in dependence on the hitch angle φ of the towed vehicle 3 determined by the first image processing module 36A. The guidance module 36C operates to determine the target route R between the current trailer position and the target position Ptar. The vehicle steering angle Θ is controlled such that a centre of rotation of the towed vehicle 3 at least substantially matches the determined target route R. The guidance module 36C implements a geometric algorithm to generate the target route R. The guidance module 36C may, for example, utilise the angular offset between the current trailer longitudinal axis X2 and the target longitudinal axis X1tar; and the lateral offset between the current trailer position and the parking destination Ptar. The guidance module 36C can optionally be configured to define a minimum radius of curvature for the first and second curves A, B to ensure that the hitch angle Φ does not equal or exceed the jack-knife angle.
The parking assist controller 4 attempts to identify the presence of any obstructions along the target route R. Typical obstructions include kerbs, walls, vehicles, etc. The parking assist controller 4 can optionally also determine terrain features. The terrain features may comprise one or more of the following set: an incline or gradient of the surface; surface roughness, for example to differentiate between a smooth surface and a rough or uneven surface. As described herein, the obstructions may be identified by the third image processing module 36D. The guidance module 36C modifies the target route R to avoid any obstructions identified by the third image processing module 36D.
The steering angle Θ of the towing vehicle 2 is controlled to maintain a travel direction of the towed vehicle 3 substantially coincident with a target trailer travel direction while reversing along said target route R. The guidance module 36C controls the steering angle Θ of the front wheels W1, W2 to guide the towed vehicle 3 along the target route R. The control algorithm for generating a control signal to guide the towing vehicle 2 along the target route R will now be described with reference to Figure 8. The towing vehicle 2 has a first longitudinal axis X1 and the towed vehicle 3 has a second longitudinal axis X2. The angular offset between the first and second longitudinal axes X1, X2 is referred to as the hitch angle Φ. During reversing, the towed vehicle 3 travels in a direction Tact corresponding to the hitch angle Φ (unless the hitch angle Φ exceeds a jack-knife angle for the towed vehicle 3).
The first image processing module 36A calculates the hitch angle Φ with reference to the target 25 and outputs a hitch angle signal to the guidance module 36C. When reversing, the guidance module 36C calculates the required steering angle θ based on the following equation:
'* ·' X < X (λ., ~ v minimaxΙα I
Where: 8t+i and 0t are the steering angles of the towing vehicle 2 at frame t+1 and t (auto steering command from the algorithm and current steering from the CAN respectively);
ΦΓθρ and Φακ are the requested and current hitch angles;
a is the maximum steering offset value; and k is a constant multiplier.
The maximum steering offset value a defines a maximum change (offset) in the steering angles within the defined time period (i.e. within the time frame t to t+1). The maximum steering offset value a thereby defines a slew rate limit. The value of the gain k can be calculated based on the relationship between θ and Φ, as shown in Figure 8. When the trailer hitch length L plus the tow bar offset of the vehicle h is equal to the vehicle wheelbase d, then the relationship between θ and Φ is one (1) for small angles and so the gain kcan be set to a value of one (1). The gain k can therefore be calculated based on the following equation:
f I 4£ ~--4
Where: L is the hitch length of the towed vehicle 3;
h is the tow bar offset of the towing vehicle 2; d is the wheelbase of the towing vehicle 2;
The gain k thereby compensates for the fact that longer trailers take longer to reach the requested hitch angle Φ.
The guidance module 36C is configured to calculate a maximum permissible hitch angle Φμαχ. If the hitch angle Φ exceeds the maximum permissible hitch angle Φμαχ (which in the present embodiment is the jack-knife angle of the towed vehicle 3), it is no longer possible to reduce the hitch angle Φ by continuing to reverse (i.e. the towed vehicle 3 has jack-knifed). If the detected hitch angle Φ is equal to or exceeds the calculated jack-knife angle, the parking assist controller 4 will advise the user to drive forward until it determines that the trailer angle δ or hitch angle Φ at the next interval is achievable while considering space and obstructions. With reference to Figure 8, the guidance module 36C calculates the maximum permissible hitch angle Φμαχ by applying the following set of equations:
φ = cos
fcr positive Θ φ = - cos
- O - + *‘i~ - I-
Where: | R is the turning radius; Θ is the steering angle of the towing vehicle 2; d is the wheelbase of the towing vehicle 2; h is the tow bar offset of the towing vehicle 2; L is the hitch length of the towed vehicle 3; |
The wheelbase d, the tow bar offset h and the maximum steering angle Θμαχ of the towing vehicle 2 are defined for the towing vehicle 2. The hitch length L of the towed vehicle 3 is entered during setup by the user (or can be determined during a calibration exercise). The guidance module 36C outputs a maximum hitch angle signal to indicate the maximum permissible hitch angle Φμαχ for the current steering angle Θ. The guidance module 36C is configured to inhibit the trailer travel direction to values which are less than the maximum permissible hitch angle Φμαχ. A minimum radius of curvature of the target route R can be defined to ensure that the trailer travel direction is less than or equal to the maximum permissible hitch angle Φμαχ.
The guidance module 36C calculates the initial required steering angle θ to achieve the desired hitch angle Φ, then iteratively calculates the required steering angle θ and hitch angle Φ. In use, the guidance module 36C is configured to output a steering control signal operative to control the steering angle θ of the front wheels W1, W2 to adjust the hitch angle Φ, thereby modifying the actual trailer travel direction. Specifically, the guidance module 36C adjusts the steering angle θ (which can be positive or negative in relation to a reference point) to increase or decrease the hitch angle Φ. The actual trailer travel direction can thereby be matched to the target trailer travel direction required to follow the target route R. The hitch angle signal is updated at least substantially in real time and the guidance module 36C makes corresponding real time adjustments to the steering control signal. The guidance module 36C controls the hitch angle Φ so that the difference between the actual trailer travel direction and the target trailer travel direction is at least substantially zero. A tolerance of +/0.5° between the actual trailer travel direction and the target trailer travel direction provides a good result.
A parking control signal S1 is output by the guidance module 36C to the EPAS 18 to adjust the steering angle Θ to control the towing vehicle 2 to match the actual travel direction of the towed vehicle 3 with the target travel direction. The controller HMI 37 can optionally output instructions to prompt the user to select the appropriate drive gear and to operate the vehicle brakes and throttle. The EPAS 18, in dependence on the parking control signal S1 output from the guidance module 36C, steers the towing vehicle 2 to guide the towed vehicle 3 along the target route R. A notification can be output by the controller HMI 37 to notify the user when the towing vehicle 2 and the towed vehicle 3 are in the target position Ptar. The parking assist controller 4 can thereby facilitate reversing the towing vehicle 2 and the towed vehicle 3 to the target position Ptar.
The present invention is applicable to a range of different types of towed vehicle 3. For example, the towed vehicle 3 can be a trailer, a caravan, a goods trailer, a flatbed trailer, a van trailer, a livestock trailer, a horsebox and so on. Similarly, the present invention is applicable to a range of different types of towing vehicles 2. For example, the towing vehicle 2 can be a motor vehicle, such as a utility vehicle, an off-road vehicle, a sports utility vehicle (SUV); or a towing engine or tractor for a semi-trailer truck.
Furthermore, although the parking assist controller 4 has been described with reference to a set of towing vehicle cameras 8, 9, 10, it will be appreciated that one or more of the cameras may be omitted. For example, the left and right towing vehicle cameras 9, 10 could be omitted. Alternatively, or in addition, more than one camera may be provided on the towed vehicle 3, for example left and right cameras may be mounted to the towed vehicle 3. In a further alternative, the towing vehicle cameras 8, 9, 10 or the rear trailer camera 15 could be used exclusively. A set of parking sensors provided on the towed vehicle 3 may provide notification to the driver of any obstructions. The hitch angle Φ could be measured directly, for example by a rotary encoder.
The parking assist controller 4 could be used to assist with the parking of a vehicle which is not being towed. The parking assist controller 4 may be configured to facilitate parking of a host vehicle. The parking assist controller 4 may be configured to identify the parking location indicator 26 generated by the parking beacon; and determine the target position
Ptar for parking the vehicle 2. The parking assist controller 4 may generate a parking assist signal for parking the vehicle 2 in dependence on the determined target position Ptar. The parking assist controller 4 may generate a target route R for parking the vehicle 2. The parking assist signal may provide partial or fully autonomous control of the vehicle 2. This parking assist controller 4 could be installed on a trailer incorporating an on-board drive mechanism, for example comprising one or more electric machines for driving the wheels of the trailer.
As described herein, the controller HMI 37 is provided in the towing vehicle 2 to control operation of the parking assist system 1. Alternatively, or in addition, the beacon HMI 32 may be configured to control operation of the parking assist system 1. The beacon HMI 32 may, for example, comprise a display screen and user input means. At least in certain embodiments the beacon HMI 32 may enable activation of a “Park Now” control function. When activated, the “Park Now” control function may initiate an automated parking sequence to cause the towing vehicle 2 to reverse the towed vehicle 3 along the target route R. The parking assist system 1 may be configured to require positive activation of the beacon HMI 32 by the user in order to continue the automated control of the towing vehicle 2. For example, the parking assist system 1 may be configured to bring the towing vehicle 2 to a stop upon determining that the user is no longer activating the beacon HMI 32. For example, if the user releases an activation button or a touch sensor provided in the beacon HMI 32, the parking assist system 1 may be configured to stop the towing vehicle 2. This control strategy may provide an additional safety requirement for the movement to cease in the event of an unsafe condition arising. The beacon HMI 32 may provide additional control functions, for example to enable a user to preview one or more target routes R, for example by scrolling or cycling through optional routes. The user may then operate the beacon HMI 32 to select one of the previewed target routes R for the towing vehicle 2 to follow. It will be understood that the control functions of the beacon HMI 32 may also be provided in the controller HMI 37 provided in the towing vehicle 2.
A parking assist system 101 in accordance with a further embodiment of the present invention in which the parking assist controller 4 is configured to control a host vehicle 102 will now be described with reference to Figures 9 and 10. Like reference numerals are used for like components, albeit incremented by 100 to aid comprehension. The host vehicle 102 may, for example be in the form of an automobile, a utility vehicle or a sports utility vehicle (SUV).
The parking assist controller 104 is operative selectively to control the host vehicle 102 to manoeuvre the host vehicle 102 to a predetermined target position Ptar. The target position Ptar defines a target location and/or a target orientation of the host vehicle 102 for parking. As described herein, the parking assist controller 104 is operative to generate a parking assist signal S1 to control the host vehicle 102 to the target position Ptar. The parking assist signal S1 provides semi-autonomous control of the host vehicle 102. In particular, the parking assist signal S1 is output to an electronic power assisted steering (EPAS) module 118 to control the steering angle θ of the front wheels W1, W2 of the host vehicle 102. The parking assist signal S1 may optionally control a torque request signal which is output to a torque generating machine, such as an internal combustion engine or an electric machine; and/or a brake torque request signal to a braking apparatus. Alternatively, the parking assist signal S1 may be output to a human machine interface (HMI), for example comprising a display screen or a head up display (HUD), to provide instructions or prompts to the driver of the host vehicle 102. The parking assist controller 104 is configured to determine the target position Ptar with reference to a parking location indicator 126 generated by the parking beacon 105.
The parking location indicator 126 in the present embodiment is projected onto the ground by at least a first projection system 128 provided in a parking beacon 105. The parking beacon 105 is unchanged from the embodiment described herein with preference to Figures 1 to 7. In particular, the first projection system 128 is configured to project the parking location indicator 126. A beacon control unit 127 is provided to control the first projection system 128 to encode a signal into the parking location indicator 126. The encoded signal may be an identification signal to facilitate detection and/or identification of the parking location indicator 126. The parking beacon 105 may optionally comprise means for sensing the host vehicle 102, such as a proximity sensor, to provide feedback to the parking assist controller 104. The parking beacon 105 may optionally comprise a camera for transmitting image data to the parking assist controller 104, for example to provide an alternate view of the region behind the host vehicle 102.
The parking assist controller 104 provided in the host vehicle 102 comprises an electronic control unit (ECU) 133 having an electronic processor 134 and a memory 135, as shown schematically in Figure 9. The processor 134 is configured to execute a set of computational instructions stored on the memory 135. The processor 134 comprises image processing means for analysing image data generated by one or more cameras disposed on the host vehicle 102. In the present embodiment, a central host vehicle camera 108 is disposed in a central position; and left and right host vehicle cameras 109, 110 are disposed in respective left and right positions on the host vehicle 102. The processor 134 is configured to receive image data from each of the host vehicle cameras 108, 109, 110. The processor 134 may be configured also to receive data from other sensors disposed on the host vehicle 102. The processor 134 may, for example, receive data from one or more of the following: ultrasonic sensors, radar systems 113A, 113B and Lidar sensors.
The processor 134 comprises an image processing module 136B for receiving image data from the host vehicle cameras 108, 109, 110. The image processing module 136B is configured to analyse the image data to identify the parking location indicator 126 generated by the parking beacon 105. In particular, the image processing module 136B processes the image data to identify any elements in the image data which change in accordance with the encoded signal introduced by the beacon control unit 127. The image processing module 136B may thereby identify the parking location indicator 126. The processor 134 determines the target position Ptar for parking the host vehicle 102 in dependence on the parking location indicator 126. In the illustrated arrangement, the light path 131 generated by the parking beacon 105 indicates a side boundary of the target position Ptar and a longitudinal extent of the target position Ptar. In a variant, the light path 131 generated by the parking beacon 105 could represent a centreline of the target position Ptar. The ends of the parking location indicator 126 represent the front and rear extremities of the target position Ptar.
The processor 134 further comprises vehicle guidance means in the form of a guidance module 136C. The guidance module 136C is provided to assist in guiding the host vehicle 102, for example to reverse the host vehicle 102 from a current position PCur to a target position Ptar. The target position Ptar may, for example, be a target parking position for the towed vehicle 303 (and optionally also the towing vehicle 302). It will be understood that the guidance module 136C could also be used to drive the host vehicle 102 into a target parking position while travelling in a forward direction. The guidance module 136C is configured to output a control signal for controlling the steering angle θ of the front wheels W1, W2 of the host vehicle 102 along a target route (or trajectory) R. The target route R is generated by the guidance module 36C to guide the host vehicle 102 from the current position PCur to the target position Ptar. The target route R is illustrated by a broken line in Figure 10. The guidance module 136C is configured to generate the target route R to guide the host vehicle 102 from its current location to the identified target position Ptar. The target route R defines a trajectory for the host vehicle 102 from the current position PCur to the target position Ptar. The target route R can comprise rectilinear and/or curved sections. The target route R is arranged coincident with a midpoint of the host vehicle 102. The current position PCur of the host vehicle 102 is monitored continuously and compared to the originally calculated target route R. Small deviations are managed within the parking assist controller 104. Larger deviations can trigger a recalculation of the target route R. If the target position Ptar becomes unachievable from the current position PCur, the user is alerted and a corrective manoeuvre is suggested (for example travel forward a short distance).
The processor 134 comprises an obstacle detection module 136D configured to identify obstacles in the trajectory of the host vehicle 102. In particular, the obstacle detection module 136D is configured to identify obstructions along the target route R of the host vehicle 102. The obstacle detection module 136D may analyse image data received from the host vehicle cameras 8,9, 10 to identify the presence/absence of obstructions, particularly along or proximal to the target route R. Alternatively, or in addition, the obstacle detection module 136D may receive detection signals from the left and right radar systems 113A, 113B to identify the presence/absence of obstructions proximal to the towing vehicle 2 and the towed vehicle 3.
The controller HMI 137 may optionally comprise input means 141 which can be operated by the user to adjust the position of the target position indicator 139 within an image output to a display screen 138. The input means 141 could comprise a touch sensitive screen and/or a rotary dial, for example. If required, the user can adjust or refine the position of the target position Ptar determined by the image processing module 136B. The controller HMI 137 could optionally also be configured to allow the user to adjust the target route R generated by the guidance module 136C.
In use, the user positions the parking beacon 105 externally of the host vehicle 102. The parking beacon 5 is typically placed on the ground. The first projection system 128 is activated to project the parking location indicator 126. The user configures the first projection system 128 such that parking location indicator 126 corresponds to a desired target position Ptar. The user may manually configure the first projection system 128 or may utilise the controller HMI 137 to set the location of the parking location indicator 126. The parking location indicator 126 is projected onto the ground and thereby provides a visible indication of the target position Ptar. In the present embodiment, the parking location indicator 126 represents a boundary of the target position Ptar.
The image processing module 136B analyses the image data from the host vehicle cameras 108, 109, 110 to identify the parking location indicator 126 generated by the parking beacon 105. In particular, the image processing module 136B processes the image data to identify the parking location indicator 26. The second image processing module 36B may optionally process the image data to identify an encoded signal introduced by the beacon control unit 127. The identification of the encoded signal may facilitate identification of the parking location indicator 26 within the image data. The image processing module 136B thereby identifies the parking location indicator 126 within the image data. The processor 134 then determines the target position Ptar for parking the host vehicle 102 based on the parking location indicator 126. The processor 134 may define the target position Ptar with reference to a target longitudinal axis X1tar. The processor 34 then generates the parking assist signal S1.
The guidance module 136C operates to determine the target route R between the current position Pcur of the host vehicle 102 and the target position Ptar. The guidance module 136C implements a geometric algorithm to generate the target route R. The guidance module 136C may, for example, utilise the angular offset between the current vehicle longitudinal axis X1 and the target longitudinal axis X1tar; and the lateral offset between the current position Pcur of the host vehicle 102 and the target position Ptar.
The parking assist controller 104 attempts to identify the presence of any obstructions along the target route R. Typical obstructions include kerbs, walls, vehicles, etc. The parking assist controller 104 can optionally also determine terrain features. The terrain features may comprise one or more of the following set: an incline or gradient of the surface; surface roughness, for example to differentiate between a smooth surface and a rough or uneven surface. As described herein, the obstructions may be identified by the obstacle detection module 136D. The guidance module 136C modifies the target route R to avoid any obstructions identified by the obstacle detection module 136D.
In use, a parking control signal S1 is output by the guidance module 136C to the EPAS 18 to adjust the steering angle θ to control the host vehicle 102 to match the actual travel direction of the host vehicle 102 with the target travel direction. The controller HMI 137 can optionally output instructions to prompt the user to select the appropriate drive gear and to operate the vehicle brakes and throttle. The EPAS 118, in dependence on the parking control signal S1 output from the guidance module 136C, steers the host vehicle 102 to follow the target route R. A notification can be output by the controller HMI 137 to notify the user when the host vehicle 102 is in the target position PTAR. The parking assist controller 104 can thereby facilitate reversing the host vehicle 102 to the target position PTAR.
A parking assist system 201 in accordance with a further embodiment of the present invention will now be described with reference to Figures 11 and 12. This embodiment is a development of the embodiment described herein with reference to Figures 1 to 7. Like reference numerals are used for like components, albeit incremented by 200 to aid comprehension.
The parking assist controller 204 in accordance with the present embodiment is configured to transmit the parking control signal S1 to control a remote vehicle 250. The remote vehicle 250 in this variant is a self-propelling vehicle, for example comprising one or more torque generating machines for outputting a traction force. The parking assist controller 204 is installed in the host vehicle 202 and is configured to identify the parking location indicator 226 generated by the parking beacon 205 in order to determine the target position Ptar. The host vehicle 202 comprises a first transceiver 245 for communicating with the remote vehicle 250. In particular, the parking assist controller 204 is configured to activate the first transceiver 245 to transmit the parking control signal S1 wirelessly to the remote vehicle 3, for example using a suitable wireless communication protocol. The host vehicle 202 may thereby directly or indirectly control operation of the remote vehicle 250. At least in certain embodiments, the host vehicle 202 may operate in a “master” capacity and the separate vehicle 250 in a “slave” capacity. This embodiment of the present invention will now be described in more detail with reference to Figure 12. Like reference numerals are used for like components.
As illustrated in Figure 12, the remote vehicle 250 comprises a self-propelling (or active) trailer. The remote vehicle 250 in the present embodiment comprises first and second torque generating machines 251, 252 arranged to drive respective left and right wheels TW1, TW2 respectively. The first and second torque generating machines 251, 252 each comprise an electric traction machine and a power inverter connected to an on-board traction battery (not shown) or other energy storage means. An electronic control unit (ECU) 253 is provided to control operation of the first and second torque generating machines 251,252. By controlling the relative speed of said first and second torque generating machines 251, 252, the ECU 253 can control the remote vehicle 250 forwards and backwards, either in a straight line or in a turning motion (proving the remote vehicle 250 with so-called tank controls). The ECU 253 is connected to a second transceiver 254 for communication wirelessly with the host vehicle 202. In use, the remote vehicle 250 may be configured to be towed behind the host vehicle 202. However, to facilitate parking of the remote vehicle 250 and/or coupling of the remote vehicle 250 to the host vehicle 202, the remote vehicle 250 may propel itself independently of the host vehicle 202. The host vehicle 202 and the remote vehicle 250 may communicate with each other, for example using a suitable wireless communication protocol.
In the present embodiment, the parking assist controller 204 is incorporated into the host vehicle 202. As illustrated in Figure 11, the host vehicle 202 comprises at least one host vehicle camera 208 configured to transmit image data to the parking assist controller 204. The image data from the at least one host vehicle camera 208 may be supplemented with image data from one or more cameras (not shown) disposed on the remote vehicle 250. The remote vehicle 250 may comprise additional sensors, for example proximity sensors which output data to the ECU 253. The parking assist controller 204 comprises a processor 234 configured to implement an image processing module 236B. The image processing module 236B analyses the image data from the host vehicle camera 208 to identify the parking location indicator 226 generated by the parking beacon 205. The operation of the image processing module 236B is the same as described herein with respect to the previous embodiment(s). The processor 234 determines the target position Ptar for parking the remote vehicle 250 based on the parking location indicator 126. The processor 234 may define the target position Ptar with reference to a target longitudinal axis X1tar.
The processor 234 provided on the host vehicle 202 comprises a guidance module 236C which determines the target route R between the current position Pcur of the remote vehicle 250 and the target position Ptar. The guidance module 236C implements a geometric algorithm to generate the target route R. The guidance module 236C may, for example, utilise the angular offset between the longitudinal axis X2 of the remote vehicle 250 and the target longitudinal axis X1tar; and the lateral offset between the current position Pcur of the remote vehicle 250 and the target position Ptar.
The parking assist controller 204 may optionally also attempt to identify the presence of any obstacles along the target route R. Typical obstacles include kerbs, walls, vehicles, etc. The parking assist controller 204 can optionally also determine terrain features. The terrain features may comprise one or more of the following set: an incline or gradient of the surface; surface roughness, for example to differentiate between a smooth surface and a rough or uneven surface. As described herein, the obstructions may be identified by the obstacle detection module 236D. The guidance module 236C modifies the target route R to avoid any obstacles identified by the third image processing module 236D.
In use, the guidance module 236C transmits the parking control signal S1, via the first transceiver 245, to the remote vehicle 250. The ECU 253 controls steering of the remote vehicle 250 in dependence on the received the parking control signal S1. As illustrated in Figure 12, the ECU 253 controls the remote vehicle 250 to follow the target route R. A notification may be output to notify the user when the remote vehicle 250 is in the target position Ptar. The parking assist controller 204 can thereby facilitate reversing the selfpropelled trailer 250 to the target position Ptar.
The operation of the parking assist controller 204 to control the remote vehicle 250 is illustrated with reference to a block diagram 73 shown in Figure 13. The processor 234 determines the current position PCur of the remote vehicle 250 (BLOCK D1). The processor 234 may, for example, analyse image data from the host vehicle camera 208 provided on the vehicle 202 to identify the current position Pcur of the remote vehicle 250. Alternatively, or in addition, the vehicle 202 may determine the current position Pcur using triangulation techniques, for example by measuring a time-of-flight of wireless signals transmitted to or from the remote vehicle 250. The processor 234 receives image data from the host vehicle camera 208 (BLOCK D2); and identifies the parking location indicator 226 (BLOCK D3). The guidance module 236C determines the target route R from the current position Pcur of the remote vehicle 250 to the target position Ptar (BLOCK D5). The processor 234 identifies obstacles along the target route R (BLOCK D6). The third image processing module 36D may optionally receive additional image data from one or more cameras provided on the host vehicle 202 and/or the remote vehicle 250 (BLOCK D7). If required, the guidance module 36C may modify the target route R to avoid any identified obstacles (BLOCK D8). It will be understood that any obstacles may be identified prior to generation of the target route R. The processor 234 generates the parking assist signal S1 (BLOCK D9). The parking assist signal S1 is transmitted to the remote vehicle 250, for example as a wireless signal (BLOCK B10).
A parking assist system 301 in accordance with a further embodiment of a parking beacon 305 in accordance with an aspect of the present invention will now be described with reference to Figures 14 to 18. The parking beacon is a development of the embodiment illustrated in Figure 3. Like reference numerals are used for like components, albeit incremented by 300 to aid comprehension. The parking beacon 305 is described with reference to a parking assist controller 304 disposed in a towing vehicle 302 coupled to a towed vehicle 303, but it will be understood that the parking beacon 305 may be used in the other embodiments described herein.
As shown in Figure 14, the parking beacon 305 comprises a beacon control unit 327 and at least a first projection system 328 for projecting a parking location indicator 326. The first projection system 328 in the present embodiment comprises a light source 329; and a controllable optical guide means 330. The optical guide means 330 is configured to define a light path 331 for the light emitted from the light source 329. The optical guide means 330 in the present embodiment comprises a mirror 330. The light source 329 in the present embodiment comprises a laser diode for emitting a laser. The light source 329 is configured to emit a beam of light along the light path 331 to generate the parking location indicator 326. The beam of light is incident on the ground and provides a visual representation of a position and/or an orientation and/or a profile of the target position Ptar. The parking beacon 305 also comprises a beacon human machine interface (HMI) 332 for controlling operation of the parking beacon 305.
The light source 329 comprises a laser diode. The mirror 330 is an optical mirror configured to reflect the light emitted from the light source 329. The mirror 330 is a scanning mirror and the orientation of the mirror 330 is adjustable dynamically to modulate the emitted light. The mirror 330 may, for example, comprise a micro-opto-electromechanical system (MOEMS). The mirror 330 in the present embodiment is pivotable about a first pivot axis X1 and a second pivot axis Z1. The first and second pivot axes X1, Z1 are disposed substantially perpendicular to each other. The mirror 330 is pivotable about said first and second axes X1, Z1 to adjust the direction of the light path 331. The beacon control unit 327 is configured to control the orientation of the mirror 330 to trace the profile of the parking location indicator 326. The beacon control unit 327 controls the mirror 330 to project the light onto the ground so that the parking location indicator 326 is visible. The beacon control unit 327 may selectively energize and de-energize the light source 329 to generate a continuous or interrupted trace. The parking location indicator 326 in the present embodiment comprises a rectangular frame corresponding to an outer perimeter of the target position Ptar.
The beacon control unit 327 is configured to control the energization of the light source 329 and the orientation of the mirror 330 to form the parking location indicator 326. By changing the orientation of the mirror 330, the light emitted by the light source 329 traces the parking location indicator 326 onto the ground, as illustrated in Figure 15. The resulting parking location indicator 326 is detectable by the parking assist controller 304 and may also provide a transient visual representation of the target position Ptar. The direction of the light path 331 may be controlled to define one or more sides of the target position Ptar.
As illustrated in Figures 16A and 16B, the parking beacon 305 in the present embodiment is configured also to project a terrain reference pattern 355 onto the ground to enable the parking assist controller 304 to assess one or more terrain features. For example, the parking assist controller 304 may assess local changes in a gradient or incline of the target position Ptar. The terrain reference pattern 355 comprises a mesh 356 composed of a plurality of rectangles 357-n having predetermined dimensions. The rectangles 357-n in the present embodiment are all substantially the same size. In a variant, the rectangles 357-n within the terrain reference pattern 355 may have different sizes, for example to facilitate determination of the orientation of the parking location indicator 326. The terrain reference pattern 355 could be generated by the light source 329. However, as illustrated in Figure 14, the terrain reference pattern 355 in the present embodiment is generated by a second projection system 358 provided in the parking beacon 305. The second projection system 358 is configured to project the terrain reference pattern 355 onto the ground. When projected onto the ground, the profile of the terrain reference pattern 355 is distorted by features and/or characteristics of the terrain. The distortion may, for example, comprise a deviation from an expected line shape, typically a straight line, and convert deviations into a contour map. The mesh 356 functions as a gradient (contour) mesh to enable determination of the terrain feature(s) and/or characteristic(s). It will be understood that the terrain reference pattern 355 may take a different form. For example, the terrain reference pattern 365 may comprise a plurality of regular or irregular geometric shapes, such as lines, curves, polygons (triangles, squares, pentagons, hexagons, etc.), circles or ellipses. The geometric shapes may have predetermined dimensions and may be arranged in a predetermined configuration, for example at a known distance apart from each other. The geometric shapes may, for example, be arranged in an array having a predetermined spacing therebetween.
As illustrated in Figures 16A and 16B, variations in the contours and profile of the ground distort the shape of the terrain reference pattern 355 projected onto the ground. If the terrain reference pattern 355 is projected onto a section of ground which is flat and substantially horizontal, the projected rectangles 357-n closely correlate to the known predetermined pattern. However, the terrain reference pattern 355 may be projected onto a section of ground which is uneven, for example comprising one or more of the following features: an inclined section, a dip or depression, a hump or mound. If the terrain reference pattern 355 is projected onto an uneven section of ground, there is less correlation between the projected rectangles 357-n and the known predetermined pattern. The profile of the terrain reference pattern 355 projected onto the ground is dependent on the profile of the section of ground, as illustrated in Figure 16B. The spacing between transverse lines making up the terrain reference pattern 355 may increase or decrease depending on the local gradient of the ground. For example, the spacing between adjacent transverse lines will increase in a region where the ground slopes downwardly away from the parking beacon 305; and the spacing between adjacent transverse lines will decrease in a region where the ground slopes upwardly away from the parking beacon 305. An incline from one side to the other (from left to right or from right to left) may have a corresponding effect on the spacing between longitudinal lines making up the terrain reference pattern 355. Alternatively, or in addition, the angular orientation of lines relative to each other within the image data may be used to derive terrain features. For example, adjacent lines which are parallel to each other (or at a known angle relative to each other) within the terrain reference pattern 355 may be distorted by the terrain features and appear inclined at an angle relative to each other within the captured image. A discontinuity or interruption in the terrain reference pattern 355 may indicate the presence of an obstacle, such as a rock or boulder. Similarly, the detection of a plurality of discontinuities or interruptions may indicate that the surface is rough or uneven.
A camera 315 disposed on the towed vehicle 303 is configured to capture image data comprising the projected terrain reference pattern 355. The parking assist controller 304 analyses the image data to determine one or more terrain features. In particular, the parking assist controller 304 is configured to analyse the image data to determine an incline of the terrain within the target position Ptar. As illustrated in Figure 15, the parking assist controller 304 in the present embodiment comprises a terrain modelling module 336E. The terrain modelling module 336E is configured to compare a stored (reference) image of the terrain reference pattern 355 with an actual image of the projected terrain reference pattern 355, as captured by the camera 315. The terrain modelling module 336E detects variations between the stored (reference) image of the terrain reference pattern 355 and the actual image captured by the camera 315 to determine the local terrain characteristics.
It will be appreciated that the profile of the rectangles 357-n is also dependent on the relative position of the camera 315. As the towed vehicle 303 approaches the target position Ptar, the shape of the rectangles 357-n may change. The terrain modelling module 336E may be configured to apply corrections to the image data to allow for any such changes in perspective. The terrain modelling module 336E may, for example, detect the beacon 305 within the image data. The dimensions of the beacon 305 can be predefined and may be used to calibrate the image data, for example by calculating a distance between the camera 315 and the terrain reference pattern 355. In the arrangement illustrated in Figure 12, the host vehicle 202 remains stationary and it is not necessary to apply corrections to account for the change in perspective.
Rather than apply a correction to the image captured by the camera 315 disposed on the towed vehicle 303, the parking beacon 305 may be provided with one or more imaging sensors, such as a camera (not shown), for capturing an image. The imaging sensor(s) may capture light in the visible spectrum or outside the visible spectrum, for example infrared or ultraviolet (UV) light. The camera provided on the parking beacon 305 should be offset from the axis of the second projection system 358, for example in a vertical direction, to facilitate detection of distortions in the projected image. The imaging sensor(s) provided on the parking beacon 305 may be configured to capture an actual image of the projected terrain reference pattern 355. The parking beacon 305 may comprise a transmitter configured to transmit the image data to the parking assist controller 304 for analysis by the terrain modelling module 336E. The parking assist controller 304 may compare the captured image data with the stored (reference) terrain reference pattern 355 to determine the terrain features. It will be understood that the perspective of the terrain reference pattern 355 would not change as the location of the imaging sensor(s) remains at least substantially fixed. In a variant, the terrain modelling module 336E could be incorporated into the parking beacon 305. The terrain modelling module 336E may be configured to compare the terrain reference pattern 355 with an actual image of the projected terrain reference pattern 355 captured by the camera (not shown) provided in the parking beacon 305. In a further variant, the terrain reference pattern 355 could be generated by providing the second projection system 358 on the towing vehicle 302 or the towed vehicle 303. A camera may optionally be provided on the towing vehicle 302 or the towed vehicle 303. The terrain modelling module 336E may operate to determine one or more terrain characteristics in the direction of travel of the towing vehicle 302 or the towed vehicle 303.
The operation of the parking assist controller 304 and the parking beacon 305 to identify terrain features in dependence on the projected terrain reference pattern 355 is illustrated in a block diagram 74 shown in Figure 17. The terrain reference pattern 355 is projected on the ground by the second projection system 358 (BLOCK E1). The camera 315 provided on the towed vehicle 303 captures image data of a scene including at least a portion of the projected terrain reference pattern 355 (BLOCK E2). The terrain modelling module 336E analyses the captured image data (BLOCK E3); and identifies an incident pattern corresponding to the projected terrain reference pattern 355 (BLOCK E4). The terrain modelling module 336E compares the incident pattern to the predefined terrain reference pattern 355 and identifies distortions therein (BLOCK E5). The terrain modelling module 336E analyses the distortions to identify one or more terrain features (BLOCK E6). The terrain modelling module 336E may generate a map of the identified terrain features.
The terrain reference pattern 355 may at least substantially correspond to the dimensions of the parking location indicator 326, for example to identify terrain features in the target position Ptar. Alternatively, the terrain reference pattern 355 may be smaller than the dimensions of the parking location indicator 326, for example to identify terrain features in one or more regions of the target position Ptar. The terrain reference pattern 355 in the present embodiment is larger than the parking location indicator 326. This arrangement enables determination of one or more terrain characteristics on at least one side of the parking location indicator 326. As illustrated in Figure 18, the terrain reference pattern 355 may extend at least partway between the target position Ptar and a current position PCur of the towed vehicle 303 (or the towing vehicle 302). In a variant of the present embodiment, the terrain modelling module 336E is configured to identify one or more terrain features along at least a portion of the target route R between the target position Ptar and the current position PCur of the towed vehicle 303. The terrain modelling module 336E receives image data comprising the projected terrain reference pattern 355. The terrain modelling module 336E processes the image data to detect changes in the incline along the target route R. The terrain modelling module 336E may determine an incline angle; and/or an incline direction of the incline.
By identifying terrain features along the target route R, a feed forward loop can be established to enable vehicle systems to anticipate associated dynamic changes as the towing vehicle 302 and/or the towed vehicle 303 progress along the target route R. For example, the identified terrain features can be used to predict instances where the path followed by the towing vehicle 302 and/or towed vehicle 303 may deviate from an expected path, for example due to side-slip when travelling over a surface having a low coefficient of friction, such as wet grass. By pre-emptively identifying terrain features which may cause this type of behaviour, the guidance module 336C may compensate when generating the target route R. In the scenario where the towed vehicle 303 is being reversed across a side slope, the towed vehicle 303 will tend to slip down the slope. The guidance module 336C may be configured to aim the towed vehicle 303 higher up the incline to correct for this lateral motion. The amount of slip will depend on the surface and also surface conditions, such as how wet the surface is. A variant of the systems described herein may derive surface features or characteristics, for example using an imaging system or wheel slip measurements to differentiate between grass and tarmac. Alternatively, an initial assumption may be made that the surface has a low coefficient of friction and, if this is not the case, applying corrections towards the end of the manoeuvre. Another related example would be if the target location was up an incline, the target route R can be modified to make the trajectory on the slope directly up rather than across by turning more sharply early on.
As illustrated in Figure 19, detected changes in incline along the target route R may be used to estimate the traction torque (Nm) required to propel the towing vehicle 302 and the towed vehicle 303 along the target route R. If, for example, the target route traverses a hollow, the path planning may choose to use a little more momentum or, at least, to be ready to apply more traction torque when required. The terrain modelling module 336E may estimate the required propulsion torque at one or more locations along the target route R, for example to maintain a constant speed. A traction torque profile may be generated for the target route R, as illustrated in Figure 19. The estimated torque may be output to an engine control unit (not shown) and/or a transmission control unit, to enable one or more vehicle parameters to be configured pre-emptively to output the estimated torque required to propel the towing vehicle
302 and the towed vehicle 303 along said route R. If the estimated incline exceeds a predetermined threshold, the terrain modelling module 336E may provide feedback to the guidance module 336C. By anticipating the traction torque requirements, corrections to the torque may be implemented pre-emptively to allow for anticipated changes as the towing vehicle 302 and/or the towed vehicle 303 traverses the at least one terrain feature. A transmission control signal may be generated to control operation of an automatic or semiautomatic transmission provided in the towing vehicle 302, for example pre-emptively to select an appropriate gear ratio.
The operation of the parking assist controller 304 pre-emptively to determine traction torque requirements in dependence on the projected terrain reference pattern 355 is illustrated in a block diagram 75 shown in Figure 20. As described herein, the guidance module 36C determines the target route R from the current position Pcur to the target position Ptar (BLOCK F1). The terrain modelling module 336E identifies one or more terrain features along the target route (BLOCK F2). The terrain modelling module 336E may, for example, determine an incline angle and/or an incline direction on the target route R. In dependence on the terrain feature(s) identified by the terrain modelling module 336E, an estimation is made of the traction torque required to drive the towing vehicle 302 and the towed vehicle
303 along the target route (BLOCK F3). A torque control strategy is generated in dependence on the estimated traction torque (BLOCK F4). The torque control strategy may, for example, comprise a traction torque profile, as illustrated in Figure 19.
The guidance module 336C may modify the route R in dependence on the identified terrain feature(s). Alternatively, or in addition, the guidance module 336C may calculate the target route R in dependence on the at least one terrain feature identified by the terrain modelling module 336E. The guidance module 336C may, for example, circumnavigate one or more detected terrain features or obstacles. The guidance module 336C may generate the target route R to avoid an incline which exceeds a predetermined incline angle. The guidance module 336C may generate the target route R such that a traversal angle (i.e. the angle at which the target route traverses an incline) reduces a predicted roll angle of the towing vehicle 302 and/or the towed vehicle 303. The term roll angle is used herein to refer to an angle about the longitudinal axis X of the towing vehicle 302 or the towed vehicle 303.
Alternatively, or in addition, the guidance module 336C may generate the target route R to ascend or descend an incline to reduce a predicted pitch angle of the towing vehicle 302 and/or the towed vehicle 302. The term roll angle is used herein to refer to an angle about the transverse axis Y of the towing vehicle 302 or the towed vehicle 303. The guidance module 336C may be configured to generate the target route R such that a predicted pitch angle of the towing vehicle 302 and/or the towed vehicle 303 remains below a predetermined pitch angle threshold. Alternatively, or in addition, the guidance module 336C may be configured to generate the target route R such that a predicted roll angle of the towing vehicle 302 and/or the towed vehicle 303 remains below a predetermined roll angle threshold.
The guidance module 336C may generate a steering control signal to steer the towing vehicle 302 along the target route R. The steering control signal may, for example, be output to an electric power assisted steering (EPAS) system. The steering control signal can include a compensatory element at least partially to correct for a predicted direction change as the towing vehicle 302 traverses the at least one terrain feature identified by the guidance module 336C. The compensatory element may at least partially correct for a predicted direction change of the towing vehicle 302, for example a side-slip or lateral movement of the towing vehicle 302 and/or the towed vehicle 303. The compensatory element may therefore correct for an anticipated direction change caused by a turning force generated by the towing vehicle 303 traversing the at least one terrain feature. It will be understood that these control strategies may be implemented to control a host vehicle irrespective of whether a towed vehicle is connected. The control strategies are applicable when travelling in a forwards direction or when reversing.
In a modified arrangement, the terrain reference pattern 355 could also function as the parking location indicator 326. For example, the terrain modelling module 336E could be configured to determine one or more terrain characteristics and also the parking location indicator 326 in dependence on the projected terrain reference pattern 355.
As described herein, the second projection system 358 is disposed in the parking beacon 305. In a variant, the second projection system 358 may be disposed in the towing vehicle 302 or the towed vehicle 303. The terrain reference pattern 355 could be projected onto the ground in front of the towing vehicle 302 or the towed vehicle 303. The terrain modelling module 336E could thereby identify one or more terrain characteristics or features in front of the towing vehicle 302 or the towed vehicle 303.
A further embodiment of a parking beacon 305 in accordance with an aspect of the present invention will now be described with reference to Figures 21A and 21B. The parking beacon 305 is a development of the embodiment described herein with reference to Figures 14 to 17. Like reference numerals are used for like components. The parking beacon 305 is described with reference to a parking assist controller 304 disposed in a towing vehicle 302 coupled to a towed vehicle 303, but it will be understood that the parking beacon 305 may be used in the other embodiments described herein.
The parking beacon 305 is configured to project a graphical representation of at least a portion of the target route R. The target route R is transmitted by the park assist controller 304 disposed in the towing vehicle 302. The beacon control unit 327 receives the target route R and controls the projection system provided in the parking beacon 305 to project a route indicator 359 representing at least a portion of the target route R. The route indicator 359 is projected onto the ground and, at least in certain embodiments, is visible to provide an indication of the current trajectory of the towed vehicle 303. As shown in Figure 21 A, the route indicator 359 in the present embodiment comprises at least one arrow 360, but it will be understood that the route indicator 359 may take different forms.
The operation of the parking assist controller 304 and the parking beacon 305 to identify terrain features in dependence on the projected route indicator 359 is illustrated in a block diagram 76 shown in Figure 23. As described herein, the guidance module 336C determines the target route R from the current position Pcur to the target position Ptar (BLOCK G1). The route indicator 359 is projected on the ground by the first projection system 328 (BLOCK G2). The camera 315 provided on the towed vehicle 303 captures image data of a scene including at least a portion of the projected route indicator 359 (BLOCK G3). The terrain modelling module 336E analyses the captured image data (BLOCK G4); and identifies an incident pattern corresponding to the projected route indicator 359 (BLOCK G5). The terrain modelling module 336E compares the incident route indicator to an image corresponding to the projected target route R and identifies distortions in the incident route indicator (BLOCK G6). The terrain modelling module 336E analyses the distortions to identify one or more terrain features (BLOCK G7). The terrain modelling module 336E may generate a map of the identified terrain features.
It has been recognised that projecting a visible warning indicator representing a current or predicted trajectory of the towing vehicle 302 and/or the towed vehicle 303 may prove useful for a person (or persons) external to the vehicle. The visible warning indicator may comprise or consist of a trajectory indicator. For example, the visible warning indicator may allow a third party to determine an expected route along which the towing vehicle 302 and/or the towed vehicle 303 will travel. The projection system described herein would be suitable for generating and projecting the visible warning indicator. The projection system could be provided in the parking beacon 305. Alternatively, or in addition, the projection system may be installed in the towing vehicle 302 and/or the towed vehicle 303. It will be understood that the projection system would be operable independently of other external lighting systems, such as the headlamps. The visible warning indicator may comprise or consist of a direction marker, such as an arrow, or other indicia suitable for representing the vehicle trajectory. It is envisaged that the visible warning indicator would be provided onto the ground proximal to the vehicle. For example, the visible warning indicator may be projected onto the ground in front of or behind the vehicle depending on the direction of travel. The visible warning indicator may be projected behind the vehicle during a reversing manoeuvre. The visible warning indicator may be generated in dependence on a steering angle and/or a target route R. The visible warning indicator may also comprise a vehicle speed indicator.
The parking assist controller 304 in this embodiment may determine an expected profile of the route indicator 359 in dependence on the target route R. However, the profile of the route indicator 359 when projected varies depending on the profile of the ground. By way of example, the route indicator 359 is illustrated as being projected onto an even, substantially horizontal surface in Figure 21 A. However, if the same route indicator 359 is projected onto an uneven surface, the projection is distorted in dependence on the terrain characteristics. In accordance with an aspect of the present invention, the parking assist controller 304 is configured to compare an actual image captured by the camera 315 provided on the towed vehicle 303 with an expected profile of the route indicator 359 determined in dependence on the determined target route R. By way of example, the route indicator 359 is illustrated in Figure 21B projected onto an uneven surface comprising an inclined region. The parking assist controller 304 may be configured to determine one or more terrain features in dependence on the distortion of the route indicator 359. For example, the parking assist controller 304 may determine an incline or gradient of the surface or detect one or more obstacles along the target route R. It will be understood that this technique may be used in conjunction with the other techniques described herein, for example the projection of the terrain reference pattern 355 described with reference to Figures 14 to 16.
A variant of the parking beacon 305 described herein with reference to Figures 21A and 21B is shown in Figures 22A and 22B. In this variant, the route indicator 359 comprises first and second path lines 361-1, 361-2 representing the target route R of the vehicle. The first and second path lines 361-1,361-2 in the present embodiment correspond to a predicted path of the respective first and second trailer wheels TW1, TW2. At least in certain embodiments, projecting the path lines 361 may facilitate identification of obstacles directly in the path of the trailer wheels TW1, TW2 as it travels along the target route R.
The parking assist controller 304 may be configured to compare an actual image captured by the camera 315 provided on the towed vehicle 303 with an expected profile of the first and second path lines 361-1, 361-2 determined in dependence on the determined target route R. With reference to Figure 22B, the first and second path lines 361-1, 361-2 are projected onto an uneven surface comprising an inclined region. The parking assist controller 304 may be configured to determine one or more terrain features in dependence on the distortion of the first and second path lines 361-1, 361-2. For example, the parking assist controller 304 may determine an incline or gradient of the surface or detect one or more obstacles present along the target route R. Alternatively, or in addition, the parking assist controller 304 may determine one or more terrain features by comparing the first and second path lines 361-1,361-2. For example, the trajectory of the first and second path lines 361-1, 361-2 may be compared to determine gradient changes. If the first and second path lines 361-1, 361-2 are not parallel when projected onto the ground, the parking assist controller 304 may determine that there is a slope or gradient. It will be understood that the parking assist controller 304 may be provided in the towing vehicle 302 or the parking beacon 305.
The towed vehicle 303 may be subject to lateral movements, such as side slipping, depending on terrain features. These lateral movements may, for example, result from the towed vehicle 303 traversing an uneven or sloping surface. The parking assist controller 304 may be configured to modify the target route R in dependence on detected terrain features, such as a detected gradient or slope. The parking assist controller 304 may determine the target route R to steer the towed vehicle 303 up an incline or slope to counteract side slip. The parking assist controller 304 may determine the target route R so as to avoid holes. If the target route R cannot avoid a hole or an uneven surface, the parking assist controller 304 may be configured to counteract a turning moment of the towed vehicle 303 as a trailer wheel TW1, TW2 traverses the hole or the uneven surface. The parking assist controller 304 may be configured to modify the target route R in dependence on detected gradient changes. For example, the parking assist controller 304 may modify the target route R to ensure that the gradient in a transverse direction and/or a longitudinal direction is below a predetermined gradient threshold. Alternatively, or in addition, the parking assist controller 304 may modify the parking assist signal S1 to compensate for changes in the trajectory of the towed vehicle 303 resulting from gradient changes. For example, the parking assist signal S1 may be modified to compensate for changes in the gradient under the first and second trailer wheels TW1, TW2. The target route R could be determined to compensate for side-slip, for example when traversing an inclined surface. The parking assist controller 304 may, for example, determine the target route R so as to steer up a slope to counteract side slip.
At least some of the embodiments described herein propose that the parking assist controller for generating the target route R is provided in the towing vehicle. However, at least in certain embodiments, the parking assist controller may be provided in the parking beacon. This implementation has the advantage that the parking assist system can more readily be retrofitted, for example for use with any vehicle which is self-propelling. In the case of a self-propelling trailer, a wireless camera may be mounted to the trailer to determine an angular orientation in relation to the beacon. A remote control may be incorporated into the parking beacon for controlling the trailer. Alternatively, or in addition, a beacon camera can be provided to identify the trailer and to determine its relative position and/or orientation. It will be understood that these control strategies may also be used to control a towing vehicle.
A schematic representation of the parking beacon 305 in accordance with a further aspect of the present invention is shown in Figure 24. The beacon HMI 332 comprises a display screen 366, a user input means 367, a beacon camera 368 and a transceiver 369. The user input means 367 comprises one or more control devices, such as a mechanical switch, a button, a knob, a capacitive switch or a resistive switch. Alternatively, or in addition, the user input means 367 may be combined with the display screen 366, for example in the form of a touch-sensitive display. The beacon control unit 327 comprises an image processor for processing the image data captured by the beacon camera 368 to determine the relative position and/or orientation of the towed vehicle 303. The beacon control unit 327 is configured to output image data captured by the beacon camera 368 to the display screen 366. The beacon control unit 327 is configured to augment the image data from the beacon camera 368 with a parking location indicator 326 and optionally also a graphical representation of a target route R. The resulting augmented image may help a user to visualise the target route R and/or the parking location indicator 326. The user input means 367 may be operable to adjust the position of the parking location indicator 326 and/or to adjust the target route R. Thus, the beacon HMI 332 is configured to generate an augmented image which can be manipulated by a user to modify the target route R. Alternatively, or in addition, the position and/or orientation of the parking location indicator 326 may be manipulated by the user. The target route R and the parking location indicator 326 may be displayed as an overlay on the display image output to the display screen 366. The user validates the inputs when the desired the parking location indicator 326 has been entered. Alternatively, or in addition, the control functions of the beacon HMI 332 may be provided in a controller HMI provided in the towing vehicle 302.
The parking assist controller 304 is configured to generate the target route R in dependence on the determined position and/or orientation of the towed vehicle 303. The parking beacon 305 may optionally communicate with the towing vehicle 302 via the transceiver 369, for example to output the parking assist signal S1. By correlating changes in the orientation of the towed vehicle 303 (which may be determined using the aforementioned image processing techniques) with a known steering angle of the towing vehicle 302, the beacon control unit 327 may optionally determine the relative orientation of the towing vehicle 302 using reverse kinematics. The parking assist controller 304 may be provided in the towing vehicle 302. Alternatively, the parking assist controller 304 may be provided in the parking beacon 305, for example integrated into the beacon control unit 327. The beacon control unit 327 is configured to output the parking assist signal S1 to the transceiver 369 for wireless transmission to the towing vehicle 302.
The operation of the beacon control unit 327 to generate an augmented image 380 for output to the display screen 366 will now be described with reference to the Figures 25, 26 and 27. A typical parking scenario is represented by way of example in a scene 381 in Figure 25. The towing vehicle 302 and the towed vehicle 303 are positioned proximal to a desired parking location. In the illustrated example, the towed vehicle 303 is to be parked next to a mains electrical connection 382 whilst avoiding an obstacle 383. The mains electrical connection 382 and the obstacle 383 are both in a fixed location. The parking beacon 305 is positioned by the user on the ground proximal to the desired parking location. The parking beacon 305 is oriented such that the beacon camera 368 is directed towards the desired parking location and preferably also the towed vehicle 303. The resulting image data captured by the beacon camera 368 represents the scene 381, including the towed vehicle 303, the mains electrical connection 382 and the obstacle 383. The image data is output to the display screen 366 and displayed as an image 340, as illustrated in Figure 26. If necessary, the user may adjust the orientation of the parking beacon 305 such that the beacon camera 368 captures the desired scene elements, such as the desired parking location.
In accordance with an aspect of the present invention, the parking assist controller 304 is configured to augment the image data generated by the beacon camera 368. In particular, the parking assist controller 304 is configured to overlay graphics and information onto the image 340 to generate the augmented image 380. A first augmented image 380 is shown in Figure 26 incorporating the image 340 captured by the parking beacon 305 in the scene illustrated in Figure 25. A target position indicator 339 is overlaid onto the image 340 to provide a graphical representation of the target position Ptar relative to other features of the scene 381. As shown in Figure 26, the target position indicator 339 is scaled at least substantially to match a footprint of the towed vehicle 303 (and optionally also the towing vehicle 302). In the illustrated example, the target position indicator 339 comprises a rectangle which provides an approximate representation of the external profile of the towed vehicle 303. It will be appreciated that the target position indicator 339 may have a different profile, for example to represent more closely the actual profile of the towed vehicle 303 in plan form. The target position indicator 339 in the present embodiment also comprises a door marker 342 and an electrical connector marker 343. The door marker 342 corresponds to a door 384 provided on the towed vehicle 303; and the electrical connector marker 343 corresponds to an electrical connector 385 provided on the towed vehicle 303.
An initial position of the target position indicator 339 within the image 340 could be determined automatically, for example by analysing the image data output from the beacon camera 368. Alternatively, the user may specify the initial position of the target position Ptar. If required, the user may adjust the position of the target position Ptar. For example, the user may actuate the user input means 367 to change the position and/or orientation of the target position indicator 339 within the augmented image 380. When satisfied that the position and/or orientation of the target position indicator 339 corresponds to the position and/or orientation of the target position Ptar, the user validates the entry.
As illustrated in Figure 26, the parking assist controller 304 may be configured to identify a current position PCur of the towed vehicle 303 (and optionally also the towing vehicle 304). The parking assist controller 304 may analyse the image 340 to identify the towed vehicle 303, for example using appropriate image processing techniques. Alternatively, the user may identify the towed vehicle 303 within the image 340. In a first operation, the user may position the target position indicator 339 to identify the current position PCur (and optionally also a current orientation) of the towed vehicle 303. In a second operation, the user may position the target position indicator 339 to identify the target position Ptar (and optionally also a current orientation) of the towed vehicle 303. Other techniques may be utilised to identify the current position PCur and the target position Ptar of the towed vehicle 303.
The parking assist controller 304 is configured to calculate the target route R along which the towed vehicle 303 is to travel from the current position PCur to the target position Ptar. The target route R may be calculated at least substantially in real time. As shown in Figure 25, a target route indicator 386 is overlaid onto the image 340 to represent the target route R. The target route R may be calculated and displayed at least substantially in real time. This real time display of the target route indicator 386 may help the user to visualise changes to the target route R, for example in dependence on changes in the position and/or orientation of the target position Ptar. Alternatively, the target route R may be calculated and the target route indicator 386 displayed only after the user has validated the target position Ptar. The parking assist controller 304 may enable the user directly to modify the target route R. The user may re-position one or more points on the target route indicator 386. In embodiments in which the user input means 367 comprises a touchscreen, the user may touch and move a point on the target route indicator 386 and/or rotate a direction vector associated with that point. The user validates the target route R which is then relayed to the towing vehicle 302. The towing vehicle 302 is controlled in dependence on the target route R to reverse the towed vehicle 303 along the target route R.
It will be understood that other features may be overlaid onto the image 340 to assist the user. By way of example, a contour mesh 387 may be provided to provide an indication of the contours of the ground. The parking assist controller 304 may process the image data to estimate terrain features, for example to identify an incline or slope. The contour mesh 387 may, for example, represent said terrain feature(s). The contour mesh 387 could be generated in dependence on the projected terrain reference pattern. For example, the incident pattern could projected onto the ground be isolated from the captured image data and used to generate the contour mesh 387. Alternatively, or in addition, the parking assist controller 304 may receive data from sensors provided on the towing vehicle 302 or the parking beacon 305, for example from an inertial measurement unit (IMU) comprising one or more accelerometers and/or gyroscopes.
The parking assist controller 304 has been described herein with particular reference to parallel parking of the towed vehicle 302 in which the longitudinal axes Xi, Xitar of the current position PCur and the target position Ptar are generally parallel to each other. It will be understood that the parking assist controller 304 in accordance with the present invention is not limited to this mode of operation. By way of example, the parking assist controller 304 can be used when the longitudinal axes Xi, Xitar of the current position PCur and the target position Ptar are generally perpendicular to each other, as illustrated in a second augmented image shown in Figure 27.
In the embodiment described herein with reference to Figures 25 to 27, the display screen 366 is provided on the parking beacon 305. The augmented image 380 may be viewed by the user operating the parking beacon 305. It will be understood that the augmented image 380 could be displayed on other display screens. The parking beacon 305 could be configured to transmit the image data corresponding to the augmented image 380 to a receiver. For example, the augmented image 380 could be output to a display screen provided in the towing vehicle 302. Alternatively, or in addition, the augmented image 380 may be output to a general purpose computational device, such as a personal computer (for example a laptop), a tablet computer or a cellular telephone. In a further variant, the augmented image 380 may be generated using image data captured from a camera provided on the computational device. In this variant, it is envisaged that the parking beacon 305 would generate a reference identifier which could be detected by the computational device. The reference identifier could comprise an optical projection, for example an encoded signal. The target route R could be determined by the computational device. Alternatively, the target route R could be transmitted from the parking assist controller 304. Thus, the parking assist controller 304 and the display screen 366 may be provided in separate devices.
The beacon HMI 332 in accordance with the present embodiment may be used in addition to the first projection system 328 described herein. This may, for example, be desirable in bright sunshine where detection of the projected parking location indicator 326 is difficult. Alternatively, the beacon HMI 332 may be used in place of the first projection system 328. In such an arrangement, it will be appreciated that the target route R would not be projected onto the ground. The relative position of towed vehicle 303 and the parking beacon 305 may, for example, be determined using signal triangulation techniques. For example, the transceiver 369 may communicate wirelessly with one or more transceivers provided on the towed vehicle 303 and/or the towing vehicle 302. The time-of-flight of the communications may be used to determine the relative position.
In a variant, the beacon control unit 327 may manipulate the image data captured by the beacon camera 368 to generate a birds-eye view image. The towing vehicle 302 and/or the towed vehicle 303 may optionally also be overlaid onto the image and output to the display screen 366.
The parking beacon described herein is a separate device intended for positioning external to the towing vehicle and the towed vehicle. A user could conceivably forget to retrieve the parking beacon after completing a parking manoeuvre. The parking beacon could be left behind when the towing vehicle and the towed vehicle are driven away.
As shown in Figure 28, the parking assist system 305 may comprise a docking station 390 for docking the parking beacon 305. The docking station 390 is provided in a host vehicle which may be either the towing vehicle 302 or the towed vehicle 303. The docking station 390 comprises a dock 391 for receiving the parking beacon, a dock control unit 392, and a charging unit 393 for charging energy storage means, such as a battery (not shown), provided in the parking beacon 305. The dock 391 is configured to locate the parking beacon 305 in a predetermined position in the docking station 390. The dock 391 may comprise a mechanical retaining mechanism or a magnet for releasably retaining the parking beacon 305. The dock control unit 392 is configured to control operation the charging unit 393 and to communicate with the beacon control unit 327 when the parking beacon 305 is docked. The charging unit 393 comprises a electrical connectors (not shown) for establishing an electrical connection with corresponding electrical connectors provided on the parking beacon 305 when the parking beacon 305 is docked in the docking station 390. The dock control unit 392 comprises an interface 394 for communicating with the beacon control unit 327 when the parking beacon 305 is docked. The interface 394 also communicates with the host vehicle (either the towing vehicle 302 or the towed vehicle 303), for example by connection to a communication bus or network of the host vehicle.
The dock control unit 392 is configured to determine when the parking beacon 305 is docked and when the parking beacon 305 is undocked. The determination may, for example, be made in dependence on the status of a communication link established between the interface 394 and the beacon control unit 327. If the dock control unit 392 cannot establish communication with the beacon control unit 327, the dock control unit 392 determines that the parking beacon 305 is undocked. If the dock control unit 392 can establish communication with the beacon control unit 327, the dock control unit 392 determines that the parking beacon 305 is docked. Alternatively, or in addition, the dock control unit 392 may determine when the parking beacon 305 is docked or undocked by monitoring the status of the charging unit 393. For example, the dock control unit 392 may measure an electrical load across the charging unit 393 to determine when the parking beacon 305 is docked or undocked.
The dock control unit 392 is configured to generate an alert if the parking beacon 305 is undocked and a determination is made that the host vehicle is moving. The dock control unit 392 may, for example, communicate with the host vehicle to determine a vehicle reference speed or a parking brake status. If the vehicle reference speed exceeds a predetermined threshold or the parking brake is released while the parking beacon 305 is undocked, the dock control unit 392 generates an alert. Alternatively, or in addition, wireless communication between the dock control unit 392 and the parking beacon 305 may determine a separation distance. If the separation distance exceeds a predetermined threshold, the dock control unit 392 may generate an alert. The alert may, for example, be output from the dock control unit 392 to the host vehicle and output by the vehicle HMI. Thus, a user may be notified that the parking beacon 305 is not in the host vehicle.
In a further variant, the docking station 390 may comprise a locking mechanism 395 for engaging the parking beacon 305. The locking mechanism 395 may comprise a locking member for engaging a cooperating surface on the parking beacon 305. The locking mechanism 395 may to operable to prevent removal of the parking beacon 305. The locking mechanism 395 may be unlocked using a mechanical key. Alternatively, the locking mechanism 395 may be released using a digital key, for example to deactivate an electromagnetic lock.
It will be understood that the docking station 390 may be incorporated into the host vehicle during manufacture. Alternatively, the docking station 390 may be retro-fitted to the host vehicle, for example by connecting to an on-board electrical supply and optionally also the communication network or bus.
A variant of the parking assist system 301 described herein with reference to Figures 14 to 28 will now be described with reference to Figure 29. Like reference numerals are used for like components.
The parking assist system 301 utilises image processing techniques to identify the towed vehicle 303 in the image data captured by the beacon camera 368. To facilitate robust identification of the towed vehicle 303, at least one target 388 may be provided on the towed vehicle 303. At least in certain embodiments, the at least one target 388 is identifiable using image processing techniques. The at least one target 388 is configured to generate a unique pattern identifiable by analysing the image data captured by the beacon camera 368 (to differentiate from possible specular reflections) to enable determination of the polar location of the towed vehicle 303 relative to the beacon 305.
The at least one target 388 may, for example, be provided on a rearward-facing (back) surface of the towed vehicle 303. Alternatively, or in addition, a target may be provided on one or both opposing sides of the towed vehicle 303. The configuration of the at least one target 388 may be different on different surfaces of the towed vehicle 303. For example, the targets 388 provided on each surface of the towed vehicle 303 may have a unique pattern to facilitate differentiation between the surfaces. Each target 388 may, for example, comprise a reflective element. The at least one target 388 may each comprise a reflective lens or a reflective material suitable for returning light in the direction of transmission. A plurality of targets 388 may be arranged on a substrate or backing. The targets 388 may have a defined spatial separation and/or angular separation to facilitate identification. The targets 388 could optionally be coloured, for example comprising a coloured reflective lens.
In the arrangement illustrated in Figure 29, one (1) target 388 is provided on a rear surface of the towed vehicle 303. The target 388 comprises a reflective element configured to reflect light emitted from a light source provided in the parking beacon 305. The light source could be configured to transmit an encoded light pattern to facilitate identification of the target 388. The encoded light pattern may, for example, comprise a predetermined pattern or signal which is identifiable in the light reflected by the target 388. The image captured by the beacon camera 368 may be analysed to identify an image element (or plurality of image elements) which changes in accordance with the predetermined encoded pattern. The encoded pattern may comprise scan times associated with a particular optical guide means 330, and the image analysis may be configured to identify this pattern. A plurality of targets 388 may be provided in a known configuration (for example in a known geometric configuration) to facilitate filtering of the image data, for example to reduce or remove image noise caused by light reflecting off other surfaces.
The parking beacon 305 could be provided with a dedicated light source for transmitting light for reflection off of the target(s) 388. In the present embodiment, however, it is envisaged that the first projection system 328 is configured to scan a region where the towed vehicle 303 is likely or expected to be situated. As described herein, the first projection system 328 may comprise a light source in the form of a laser. The first projection system 328 may, for example, be configured to rotate the optical guide means 330 about the transverse axis Y1 so as to scan the beam of light in a vertical plane, as illustrated in Figure 29. The beam of light emitted from the first projection system 328 may be scanned across as a vertical fan or scanned in a raster pattern to locate the target 388. At least some of the light emitted from the first projection system 328 which is incident on the target 388 is reflected back towards the beacon 305. The reflected light is detected by processing the image data generated by the beacon camera 368, thereby enabling identification of the target 388. Using appropriate image processing techniques, the location and orientation of the target 388 may be determined. Once the target 388 is located, the optical guide means 330 may be controlled to scan the light over a reduced or limited area in the region of a last known position. Alternatively, or in addition, the approximate location of the target 388 may be determined using the beacon camera 368. The approximate location may be used to optimise the scanning performed by the optical guide means 330. By identifying the target 388, the image data may be processed to identify the location of the towed vehicle 303. The orientation of the towed vehicle 303 may be determined by tracking its movement.
A plurality of the targets 388 may be arranged in a predefined pattern, for example having a predefined angular separation between the targets 388 and/or a predefined spatial separation between the targets 388. The predefined pattern may be a geometric pattern. When the targets 388 are scanned, for example by scanning the laser emitted by the first projection system 328, the returned pulses will have a time separation which is dependent on the angular and/or spatial separation of the targets 388. The angular separation of the targets 388 and the distance to the targets 388 can be determined with reference to the angular scan speed of the optical guide means 330. The polar position of the towed vehicle 303 and the distance from the beacon 305 to the towed vehicle 303 may thereby be determined. Thus, the first projection system 328 and the beacon camera 368 may be used to determine the relative location and orientation of the towed vehicle 303.
A first set comprising one or more of the targets 388 may be provided on a first surface of the towed vehicle 303, for example provided on the rear of the towed vehicle 303. A second set comprising one or more of the targets 388 may be provided on a second surface of the towed vehicle 303, for example on a side of the towed vehicle 303. The second set of targets 388 may be oriented orthogonally relative to the first set of targets 388. The provision of first and second sets of the targets 388 may help to reduce errors in determining the orientation of the towed vehicle 303. Alternatively, any such orientation errors may be identified by analysing changes in their relative position in dependence on the trajectory of the towed vehicle 303 when it moves. A correction may be applied to correct any identified orientation errors to determine the orientation of the towed vehicle 303.
A target 388 may optionally be applied to the towed vehicle 303 to identify particular features, for example to identify the door 384; and/or the electrical connector 385.
The concept of applying at least one target 388 to the towed vehicle 303 may be employed in the other embodiments described herein. For example, at least one target 388 may be provided on the remote vehicle 250 to facilitate identification thereof. Alternatively, or in addition, the beacon 305 may be provided with at least one target 388 to facilitate identification thereof. Rather than apply a separate target 388, the concept may make use of existing reflectors provided on the towing vehicle 302, the towed vehicle 303 or the remote vehicle 305. For example, the techniques described herein may seek to identify light reflected by a reflector associated with a lamp or light provided on the vehicle. The technique, may for example, seeks to identify light reflected from one or more of the following set: a reversing light, a tail light, a side light, a turning indicator and a head lamp. The different colour of the light reflected by the different reflectors may be used to refine the location and/or orientation of the vehicle.
It will be appreciated that various modifications may be made to the embodiment(s) described herein without departing from the scope of the appended claims.
The parking assist controller 104 has been described with reference to identifying the parking location indicator 126 by analysing the image data from on-board camera(s). It will be understood that other techniques may be used to determine the location of the parking beacon 105. For example, signal triangulation could be used to determine the location of the parking beacon 105. The towing vehicle 102 and/or the towing vehicle 103 may each comprise one or more transceivers for communicating with the parking beacon 105. By measuring the time-of-flight for communications with the parking beacon 105, the parking assist controller 104 may determine the location of the parking beacon 105. A user may use one or more of said parking beacon 105 to mark a target position. The parking beacon 105 may operatively control the optical system described herein to provide a visible representation of the target position.
The parking beacon 105 may be configured to project the target path R onto the ground. For example, the parking assist controller 104 may transmit the target path R to the parking beacon 105. The parking beacon 105 may project the target path R onto the ground to enable the user to visualise the proposed route and optionally also to confirm the route. The light source 29 and the mirror 30 may be controlled to display the target path R. By visually projecting the target path R onto the ground, the user may consider aspects like ground condition and/or inclination that may be less evident when reviewing a video image output to a screen.
The systems and methods described herein may be used in conjunction with a plurality of the parking beacons. For example, the parking beacons may be arranged to form a substantially continuous chain alongside the target route.
Claims (26)
1. A terrain feature analysis system for use with a vehicle, the terrain feature analysis system comprising a processor configured to:
receive image data from one or more image sensors, the image data comprising an image of a section of ground onto which a predetermined reference pattern is projected;
process the image data to identify an incident pattern generated by the projection of the predetermined reference pattern onto the section of ground;
identify distortions in the incident pattern caused by projecting the predetermined reference pattern onto the section of ground; and identify one or more terrain features in said section of ground in dependence on the identified distortions.
2. A terrain feature analysis system as claimed in claim 1, wherein identifying distortions in said incident pattern comprises comparing the incident pattern derived from the image data with the predetermined reference pattern.
3. A terrain feature analysis system as claimed in claim 1 or claim 2, wherein the processor may be configured to identify one or more of the following distortions: compression of the predetermined reference pattern, extension of the predetermined reference pattern; skewing of the predetermined reference pattern; discontinuities in the predetermined reference pattern.
4. A terrain feature analysis system as claimed in any one of the preceding claims, wherein the predetermined reference pattern comprises a geometric pattern.
5. A terrain feature analysis system as claimed in claim 4, wherein the geometric pattern comprises a mesh.
6. A terrain feature analysis system as claimed in any one of the preceding claims, wherein the one or more terrain features comprises one or more of the following: an incline; a depression; a mound; surface roughness; and an object.
7. A terrain feature analysis system as claimed in any one of the preceding claims, wherein the predetermined reference pattern is encoded and identifying the incident pattern comprising identifying encoded image data.
8. A terrain feature analysis system as claimed in any one of the preceding claims, wherein the predetermined reference pattern is generated by a projection system disposed in a beacon.
9. A terrain feature analysis system as claimed in claim 8, wherein the image is captured by one or more image sensors disposed in a vehicle.
10. A vehicle comprising a terrain feature analysis system as claimed in any one of the preceding claims.
11. A beacon comprising a projection system for projecting a predetermined reference pattern onto a section of ground.
12. A beacon as claimed in claim 11 comprising one or more light sources for emitting light to generate the predetermined reference pattern.
13. A beacon as claimed in claim 12, wherein at least one of the one or more light sources is configured to encode the predetermined reference pattern.
14. A beacon as claimed in claim 12 or claim 13, wherein said one or more light sources comprise a laser.
15. A beacon as claimed in any one of claims 12, 13 or 14 comprising a controllable mirror for reflecting light emitted from said one or more light sources to generate said predetermined reference pattern.
16. A beacon as claimed in any one of claims 11 to 15, wherein said beacon is portable.
17. A beacon as claimed in any one of claims 11 to 16, wherein the projection system is configured to project a parking location indicator and/or a vehicle trajectory indicator.
18. A method of analysing terrain features for use with a vehicle, the method comprising:
projecting a predetermined reference pattern onto a section of ground;
capturing an image of the section of ground onto which the predetermined reference pattern is projected;
analysing the image to identify an incident pattern generated by the projection of the predetermined reference pattern onto the section of ground;
identifying distortions in the incident pattern caused by projecting the predetermined reference pattern onto the section of ground; and identifying one or more terrain features in said section of ground in dependence on the identified distortions.
19. A method as claimed in claim 18, wherein identifying distortions in said incident pattern comprises comparing the incident pattern derived from the image data with the predetermined reference pattern.
20. A method as claimed in claim 18 or claim 19 comprising identifying one or more of the following distortions: compression of the predetermined reference pattern, extension of the predetermined reference pattern; skewing of the predetermined reference pattern; discontinuities in the predetermined reference pattern.
21. A method as claimed in any one of claims 18, 19 or 20, wherein the predetermined reference pattern comprises a geometric pattern.
22. A method as claimed in any one of claims 18 to 21, wherein the terrain reference pattern comprises a mesh.
23. A method as claimed in any one of claims 18 to 22, wherein the one or more terrain features comprises one or more of the following: an incline; a depression; a mound; surface roughness; and an object.
24. A method as claimed in any one of claims 18 to 23, wherein the predetermined reference pattern is generated by a projection system disposed in a beacon.
25. A method as claimed in claim 24, wherein the image is captured by one or more image sensors disposed in a vehicle.
26. A non-transitory computer-readable medium having a set of instructions stored therein which, when executed, cause a processor to perform the method claimed in any one of claims 18 to 25.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
GB1719707.0A GB2568751A (en) | 2017-11-28 | 2017-11-28 | Terrain analysis apparatus and method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
GB1719707.0A GB2568751A (en) | 2017-11-28 | 2017-11-28 | Terrain analysis apparatus and method |
Publications (2)
Publication Number | Publication Date |
---|---|
GB201719707D0 GB201719707D0 (en) | 2018-01-10 |
GB2568751A true GB2568751A (en) | 2019-05-29 |
Family
ID=60950533
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
GB1719707.0A Withdrawn GB2568751A (en) | 2017-11-28 | 2017-11-28 | Terrain analysis apparatus and method |
Country Status (1)
Country | Link |
---|---|
GB (1) | GB2568751A (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113311408A (en) * | 2021-07-07 | 2021-08-27 | 中国地质大学(武汉) | Radiation correction method and device for hyperspectral laser radar |
WO2023001819A1 (en) * | 2021-07-20 | 2023-01-26 | Robert Bosch Gmbh | Method for aligned parking of a trailer |
IT202200000518A1 (en) * | 2022-01-14 | 2023-07-14 | Toyota Mat Handling Manufacturing Italy S P A | TRACTOR TROLLEY. |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP7395511B2 (en) * | 2019-01-16 | 2023-12-11 | 株式会社小糸製作所 | Imaging device, its processing unit, vehicle lighting, vehicle, sensing method |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5343284A (en) * | 1990-10-24 | 1994-08-30 | Kaman Aerospace Corporation | Imaging lidar system employing bistatic operation |
-
2017
- 2017-11-28 GB GB1719707.0A patent/GB2568751A/en not_active Withdrawn
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5343284A (en) * | 1990-10-24 | 1994-08-30 | Kaman Aerospace Corporation | Imaging lidar system employing bistatic operation |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113311408A (en) * | 2021-07-07 | 2021-08-27 | 中国地质大学(武汉) | Radiation correction method and device for hyperspectral laser radar |
WO2023001819A1 (en) * | 2021-07-20 | 2023-01-26 | Robert Bosch Gmbh | Method for aligned parking of a trailer |
IT202200000518A1 (en) * | 2022-01-14 | 2023-07-14 | Toyota Mat Handling Manufacturing Italy S P A | TRACTOR TROLLEY. |
EP4212820A1 (en) * | 2022-01-14 | 2023-07-19 | TOYOTA MATERIAL HANDLING MANUFACTURING ITALY S.p.A | Tractor-trailer |
Also Published As
Publication number | Publication date |
---|---|
GB201719707D0 (en) | 2018-01-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20190166338A1 (en) | Projection apparatus | |
US20190161084A1 (en) | Vehicle control apparatus and method | |
US20190161118A1 (en) | Parking assist method and apparatus | |
WO2019105665A1 (en) | Parking assist method and apparatus | |
US20190162545A1 (en) | Imaging apparatus and method | |
US20190161121A1 (en) | Parking assist method and apparatus | |
US20190161119A1 (en) | Vehicle parking apparatus | |
US20190164427A1 (en) | Docking apparatus | |
US12103342B2 (en) | Autonomous detection of and backing to trailer kingpin | |
GB2568751A (en) | Terrain analysis apparatus and method | |
US20220048497A1 (en) | System and methods for autonomously backing a vehicle to a trailer | |
GB2568750A (en) | Terrain analysis apparatus and method | |
US11951790B2 (en) | Systems and methods for determining a height of an object above a vehicle | |
CA3101729A1 (en) | Selection of environment sensors for autonomous vehicle maneuvering | |
CA3101732A1 (en) | Systems and methods for automatically updating a model of vehicle turning dynamics | |
GB2568883A (en) | Projection apparatus | |
GB2568752A (en) | Vehicle position identification method and apparatus | |
CA3102216A1 (en) | Systems and methods for autonomously backing a vehicle to a dock |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WAP | Application withdrawn, taken to be withdrawn or refused ** after publication under section 16(1) |