WO2013031161A1 - 運転支援装置 - Google Patents
運転支援装置 Download PDFInfo
- Publication number
- WO2013031161A1 WO2013031161A1 PCT/JP2012/005321 JP2012005321W WO2013031161A1 WO 2013031161 A1 WO2013031161 A1 WO 2013031161A1 JP 2012005321 W JP2012005321 W JP 2012005321W WO 2013031161 A1 WO2013031161 A1 WO 2013031161A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- dimensional object
- imaging
- host vehicle
- sensor
- Prior art date
Links
- 238000003384 imaging method Methods 0.000 claims abstract description 118
- 238000001514 detection method Methods 0.000 claims abstract description 46
- 238000012545 processing Methods 0.000 claims abstract description 44
- 240000004050 Pentaglottis sempervirens Species 0.000 claims abstract description 28
- 235000004522 Pentaglottis sempervirens Nutrition 0.000 claims abstract description 28
- 239000002131 composite material Substances 0.000 claims description 17
- 230000002194 synthesizing effect Effects 0.000 claims description 5
- 230000003287 optical effect Effects 0.000 claims description 3
- 230000008034 disappearance Effects 0.000 abstract description 3
- 238000000034 method Methods 0.000 description 17
- 238000006243 chemical reaction Methods 0.000 description 15
- 238000010586 diagram Methods 0.000 description 14
- 230000015572 biosynthetic process Effects 0.000 description 5
- 238000003786 synthesis reaction Methods 0.000 description 5
- 239000000470 constituent Substances 0.000 description 2
- 230000010354 integration Effects 0.000 description 2
- 238000005259 measurement Methods 0.000 description 2
- 239000000203 mixture Substances 0.000 description 2
- 238000007796 conventional method Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000009434 installation Methods 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R1/00—Optical viewing arrangements; Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/20—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/22—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle
- B60R1/23—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle with a predetermined field of view
- B60R1/27—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle with a predetermined field of view providing all-round vision, e.g. using omnidirectional cameras
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
- H04N7/181—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a plurality of remote sources
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/10—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of camera system used
- B60R2300/105—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of camera system used using multiple cameras
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/30—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of image processing
- B60R2300/301—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of image processing combining image information with other obstacle sensor information, e.g. using RADAR/LIDAR/SONAR sensors for estimating risk of collision
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/30—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of image processing
- B60R2300/302—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of image processing combining image information with GPS information or vehicle data, e.g. vehicle speed, gyro, steering angle data
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/30—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of image processing
- B60R2300/304—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of image processing using merged images, e.g. merging camera image with stored images
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/60—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by monitoring and displaying vehicle exterior scenes from a transformed perspective
- B60R2300/607—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by monitoring and displaying vehicle exterior scenes from a transformed perspective from a bird's eye viewpoint
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/80—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement
- B60R2300/802—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement for monitoring and displaying vehicle exterior blind spot views
Definitions
- the present invention relates to a driving support device for notifying the possibility of collision with an obstacle when parking.
- Patent Document 1 a driving support device that supports driving by synthesizing and displaying an all-around image that shows the entire circumference of a vehicle from an overhead viewpoint using images captured by a plurality of cameras.
- the all-around image is synthesized using each camera image projected on the road surface position, the disappearance of a three-dimensional object (dead angle) occurred at the junction boundary of adjacent cameras.
- the junction boundary of each camera image is set near the four corners of the vehicle due to restrictions such as the camera mounting position, its angle of view, and pixel density.
- the four corners of the vehicle are also likely to be blind spots when viewed by the driver. For this reason, the driver may continue to drive without being able to recognize the three-dimensional object in the vicinity of the joint boundary, and the three-dimensional object may collide with the vehicle.
- An object of the present invention is to provide a driving support device that can prevent a three-dimensional object from disappearing in the immediate vicinity of a vehicle in a bird's-eye view image even though the three-dimensional object is detected.
- the detection range by the sensor is included in the angle of view of the second imaging unit, and when the sensor detects a three-dimensional object, the image processing unit is configured to detect the captured image of the second imaging unit and the first imaging
- a bird's-eye view image is created by synthesizing with the captured image of the means, and a detection range by the sensor is included in a region of the bird's-eye view image based on the image taken by the second imaging means of the bird's-eye view image.
- the driving support device of the present invention it is possible to prevent the three-dimensional object from disappearing immediately in the overhead image even though the three-dimensional object is detected, so that the driver recognizes the three-dimensional object closest to the vehicle. The effect that it becomes easy to do.
- the block diagram which shows the structure of the driving assistance apparatus in Embodiment 1 of this invention.
- the figure explaining the solid object in Embodiment 1 of this invention by an image The flowchart figure of the driving assistance process by the driving assistance apparatus in Embodiment 1 of this invention.
- FIG. 6 shows an image
- the block diagram which shows the structure of the driving assistance apparatus in Embodiment 2 of this invention.
- Embodiment 1 the driving assistance apparatus of Embodiment 1 of this invention is demonstrated, referring drawings.
- the vehicle will be described using the right handle as an example.
- the left handle the left and right are reversed.
- FIG. 1 is a block diagram showing the configuration of the driving support apparatus according to the first embodiment of the present invention.
- a driving support device 1 is configured by an image ECU (Electric Control Unit) that performs image processing, and includes a volatile memory 2, an image processing unit 3, a nonvolatile memory 4, a control unit 5, and the like. And a bus 6 connected to each other.
- the driving support device 1 is connected to the first imaging unit 7, the input unit 8, the vehicle speed sensor 9, the steering sensor 10, the gear 11, the sonar 12, the display unit 13, and the second imaging unit 14.
- the input unit 8, the vehicle speed sensor 9, the steering sensor 10, the gear 11, the sonar 12, the display unit 13, and the second imaging unit 14 may be included in the driving support device 1.
- the steering sensor 10 and the steering signal shown in FIG. 1 may be rephrased as “steering angle sensor 10” and “steering angle signal”, respectively.
- the volatile memory 2 is composed of, for example, a video memory or a RAM (Random Access Memory).
- the volatile memory 2 is connected to the first imaging means 7.
- the volatile memory 2 is connected to the second imaging means 14.
- the volatile memory 2 temporarily stores image data obtained from captured images input from the first imaging unit 7 and the second imaging unit 14 every predetermined time.
- the image data stored in the volatile memory 2 is output to the image processing means 3 via the bus 6.
- the image processing means 3 includes, for example, an ASIC (Application Specific Integrated Circuit) or a VLSI (Very Large Scale Integration).
- the image processing means 3 is connected to the display means 13.
- the image processing means 3 performs viewpoint conversion on the image data input from the volatile memory 2 and creates an overhead image on which the image data input from the nonvolatile memory 4 is superimposed at predetermined time intervals.
- the image processing means 3 may create a composite image in which normal images that are not subjected to viewpoint conversion are arranged on the overhead view image.
- the technique disclosed in International Publication No. 00/64175 can be used as the viewpoint conversion technique.
- the image processing means 3 outputs the composite image created every predetermined time to the display means 13 as a display image.
- the non-volatile memory 4 is composed of, for example, a flash memory or a ROM (Read Only Memory).
- the nonvolatile memory 4 stores various image data such as image data of the host vehicle and a data table related to a display method according to the driving situation.
- the image data stored in the non-volatile memory 4 is read by an instruction from the control means 5 and used for various image processes by the image processing means 3.
- the control means 5 is composed of, for example, a CPU (Central Processing Unit) or an LSI (Large Scale Integration).
- the control means 5 is connected to the input means 8, the vehicle speed sensor 9, the steering sensor 10, the gear 11, and the sonar 12. Based on various signals input from the input means 8, the vehicle speed sensor 9, the steering sensor 10, the gear 11, and the sonar 12, the control means 5 performs the image processing of the image processing means 3, the volatile memory 2, and the nonvolatile memory 4. Data to be read, input from the first imaging means 7 and second imaging means 14, output to the display means 13, and the like are controlled.
- the first imaging means 7 is composed of four cameras each.
- the second imaging means 14 is composed of one camera.
- the first imaging unit 7 and the second imaging unit 14 input captured images captured at predetermined time intervals to the volatile memory 2 of the driving support device 1.
- the first imaging means 7 is attached to the vehicle body so that the entire periphery of the host vehicle can be imaged.
- the 2nd imaging means 14 is installed in the left front corner of the vehicle. The mounting positions of the first imaging means 7 and the second imaging means 14 on the vehicle body will be described.
- FIG. 2 is a diagram for explaining the attachment positions of the first imaging means 7 and the second imaging means 14 to the vehicle.
- the first imaging means 7 includes a front camera 7a, a right side camera 7b, a left side camera 7c, and a rear camera 7d.
- the front camera 7a and the rear camera 7d are attached to bumpers on the front and rear of the vehicle body, for example.
- the right side camera 7b and the left side camera 7c are attached to the lower part of the left and right door mirrors of the host vehicle, for example.
- the 2nd imaging means 14 is attached to the left front corner of the own vehicle.
- the input means 8 is composed of, for example, a touch panel, a remote control, and a switch.
- the input means 8 may be provided in the display means 13 when configured by a touch panel.
- the vehicle speed sensor 9, the steering sensor 10, the gear 11, and the sonar 12 are respectively a vehicle speed signal indicating the vehicle speed of the host vehicle, a steering angle signal indicating the steering angle, a gear signal indicating the state of the shift lever, a three-dimensional object detection signal, and a distance signal. Is output to the control means 5.
- the sonar 12 is composed of eight sonars, and is attached to four locations at the four corners of the vehicle body and four locations at the front and rear of the vehicle body. The attachment position of the sonar 12 to the vehicle body will be described.
- FIG. 3 is a diagram for explaining the mounting position of the sonar 12 on the vehicle.
- the sonar 12 includes a left front corner sonar 12a, a right front corner sonar 12b, a left rear corner sonar 12c, a right rear corner sonar 12d, a left front sonar 12e, a right front sonar 12f, a left rear sonar 12g, and a right rear sonar 12h. Consists of.
- the horizontal detection ranges 16e to 16h of the left front sonar 12e, the right front sonar 12f, the left rear sonar 12g, and the right rear sonar 12h are the left front corner sonar 12a, the right front corner sonar 12b, and the left rear sonar.
- the corner sonar 12c and the right rear corner sonar 12d are set to be narrower than the detection ranges 16a to 16d on the horizontal plane. Next, the relationship between the detection range of the left front corner sonar 12a and the angle of view of the second imaging means 14 will be described.
- FIG. 4 is a diagram for explaining the angle of view of the second imaging means 14 and the detection range of the horizontal plane of the left front corner sonar 12a with images.
- the angle of view 17 of the second imaging means 14 is set to approximately 180 degrees on a horizontal plane.
- the detection range 16 a of the left front corner sonar 12 a is included in the angle of view 17 of the second imaging unit 14. That is, the entire detection range 16a of the left front corner sonar 12a is included in the angle of view 17 of the second imaging means 14.
- the second imaging means 14 is attached above the vehicle body rather than the left front corner sonar 12a.
- the detection range 16a of the left front corner sonar 12a is easily included three-dimensionally within the angle of view 17 of the second imaging means 14.
- the optical axes of the second imaging means 14 and the left front corner sonar 12a are preferably substantially equal.
- the shift of the detection range 16a of the left front corner sonar 12a within the angle of view 17 of the second imaging means 14 is reduced, and a part of the detection range 16a of the left front corner sonar 12a is outside the angle of view 17 of the second imaging means 14. The risk of overhanging can be reduced.
- the display means 13 is composed of, for example, a navigation device or a rear seat display.
- the display unit 13 displays the composite image input from the image processing unit 3.
- This synthesized image may be only an overhead image or an image in which an overhead image and a normal image are arranged in parallel. If there is a blind spot near the boundary of this overhead image, the three-dimensional object disappears. Here, the three-dimensional object in this embodiment is illustrated.
- FIG. 5 is a diagram for explaining the three-dimensional object in the present embodiment as an image.
- a color cone registered trademark
- This color cone disappears more than half in three dimensions on the overhead image, it means that there is a blind spot near the boundary of the overhead image.
- FIG. 6 is a flowchart of the driving support process by the control means 5.
- step S61 the control means 5 determines whether or not the shift lever is in a reverse state based on the gear signal input from the gear 11.
- step S61 the image processing means 3 creates an overhead image using the captured image of the first imaging means 7 acquired from the volatile memory 2 according to the command of the control means 5. Then, as shown in step S62, the display unit 13 displays the created overhead image and the rear image of the rear camera 7d acquired from the volatile memory 2 in parallel.
- step S63 the image processing unit 3 uses the captured image of the first imaging unit 7 acquired from the volatile memory 2 to generate an overhead image.
- the display means 13 displays the created overhead image and the front image of the front camera 7a acquired from the volatile memory 2 in parallel.
- step S64 the control means 5 determines from the detection result of the sonar 12 whether or not a three-dimensional object exists in the left front corner of the host vehicle. That is, the control means 5 determines whether or not the left front corner sonar 12a has detected a three-dimensional object. If NO in step S64, the process of step S61 is performed again.
- step S65 the image processing means 3 creates a new overhead image using the captured image of the first imaging means 7 and the captured image of the second imaging means 14. Then, the created overhead image is displayed on the display means 13. That is, only when the left front corner sonar 12a detects a three-dimensional object, the image processing unit 3 uses the captured images of the four cameras 7a to 7d of the first imaging unit 7 and the captured images of the second imaging unit 14. Create an overhead image. In other words, when the left front corner sonar 12a does not detect the three-dimensional object of the left front corner of the host vehicle, the image processing means 3 uses only the images taken by the four cameras 7a to 7d of the first imaging means 7.
- step S66 the control means 5 determines whether or not the traveling direction of the host vehicle is moving forward. At this time, the control means 5 specifies the traveling direction of the host vehicle based on the gear signal input from the gear 11. That is, the control means 5 determines whether or not the shift lever is the front from the gear signal.
- step S67 the control means 5 replaces the rear image displayed in parallel on the display means 13 in step S62 or the front image displayed in parallel on the display means 13 in step S63.
- the captured image of the second imaging unit 14 is displayed on the display unit 13.
- step S68 the control means 5 determines whether or not the end of the driving support mode is detected. In the case of YES at step S68, the control means 5 ends the driving support process. For example, when the end of the driving support mode is input from the input unit 8, the control unit 5 ends the driving support process. On the other hand, if NO in step S68, the control means 5 performs the process of step S61 again.
- FIG. 7 is a diagram for explaining the bird's-eye view image created in step S65 as an image.
- the image processing means 3 uses a bird's-eye view image using images taken by the four cameras 7a to 7d of the first imaging means 7. 40a is created. An image 21 of the host vehicle is superimposed on the center of the overhead image 40a.
- the regions 22 to 25 of the bird's-eye view image 40a correspond to viewpoint conversion images of images captured by the front camera 7a, the right side camera 7b, the left side camera 7c, and the rear camera 7d, respectively.
- the joint surfaces of the regions 22 to 25 are represented as composite boundaries 31 to 34.
- the left front corner sonar 12a detects the three-dimensional object 41 in the vicinity of the host vehicle, a blind spot occurs at the composite boundary 34 if the overhead image using the images captured by the four cameras 7a to 7d of the first imaging means 7 is used, and the three-dimensional object is generated. 41 disappears.
- the image processing means 3 adds the three-dimensional object in addition to the picked-up images of the four cameras 7a to 7d of the first image pickup means 7, as shown on the right side of FIG.
- An overhead image 40b is created using the captured image of the second imaging means 14 that captures the left front corner from which 41 is detected.
- the regions 22 to 26 of the bird's-eye view image 40b correspond to the viewpoint conversion images of the images captured by the front camera 7a, the right side camera 7b, the left side camera 7c, the rear camera 7d, and the second imaging unit 14, respectively.
- the composite boundary 35 between the front camera 7a and the second image pickup means 14 and the composite boundary 36 between the left camera 7d and the second image pickup means 14 are such that the area 26 of the viewpoint conversion image of the second image pickup means 14 is the left front corner sonar 12a. It is set to a position that can include the detection range. In other words, the composite boundaries 35 and 36 are set outside the detection range of the left front corner sonar 12a. Accordingly, the three-dimensional object 41 detected by the left front corner sonar 12a does not disappear in the vicinity of the synthesis boundaries 35 and 36 in the overhead image 40b, and the visibility of the three-dimensional object 41 in the region 26 is maintained.
- the driver can visually recognize the three-dimensional object 41 without feeling uncomfortable.
- the three-dimensional object 41 falls down in a direction away from the vehicle in a radial direction, the position and direction of the three-dimensional object 41 can be intuitively grasped.
- the synthesis boundaries 35 and 36 are set as close as possible to the angle of view of the second imaging means 14.
- the composite boundaries 35 and 36 are made substantially equal to the angle of view of the second image pickup means 14, a reflection outside the original angle of view will occur when the second image pickup means 14 deviates from the mounting position. .
- the composite boundaries 35 and 36 are set at positions on the inner side of several degrees to ten degrees with respect to the angle of view of the second imaging unit 14.
- FIG. 8 is a diagram for explaining a blind spot area generated in the overhead image of FIG. 7 by an image.
- FIG. 8 shows a normal overhead image 50a created by images taken by the four cameras 7a to 7d of the first imaging means 7.
- the right side of FIG. 8 shows the overhead image 50b created in step S65.
- a certain range from the host vehicle image 51 is set as the blind spot measurement region 52. This certain range is set to about several tens of centimeters. For example, it is set to 50 cm. This indicates the distance that the host vehicle can travel at a creep of about 3 km / h and can be stopped by a sudden brake.
- the detection range 53 of the left front corner sonar 12a is the front camera 7a. Partially overlaps with the blind spot area 54 generated by the composite boundary 34 between the left camera 7c and the left camera 7c. If the detection range 53 of the left front corner sonar 12a overlaps with the blind spot area 54, the three-dimensional object 41 disappears within the overlap range. Therefore, in the bird's-eye view image 50b created in step S65, the composite boundaries 35 and 36 are moved away from the detection range 53 of the left front corner sonar 12a by using the captured image of the second imaging means 14. Thereby, as shown in the right side of FIG.
- the detection range 53 of the left front corner sonar 12a is a blind spot caused by the composite boundary 35 of the second imaging means 14 and the front camera 7a.
- the area 55 and the blind spot area 56 caused by the composite boundary 36 between the second imaging means 14 and the left camera 7d are separated from each other and do not overlap.
- the three-dimensional object existing in the detection range of the front corner sonar 12a is displayed on the display means 13 without disappearing on the overhead image 50b.
- FIG. 9 is a diagram for explaining the boundary of the blind spot area of FIG. 8 with an image.
- FIG. 9 will be described with an example of the bird's-eye view image 50a on the left side of FIG.
- the boundary line 54a shown in the upper left of FIG. 9 shows the left boundary of the blind spot area 54 in FIG. That is, the boundary line 54 a indicates the outer edge of the blind spot measurement region 52.
- the boundary line 54b shown in the lower left part of FIG. 9 shows the lower boundary of the blind spot area 54 in FIG.
- the boundary line 54 b indicates a boundary where a blind spot occurs when the three-dimensional object 41 is present in the region 25.
- the three-dimensional object 41 is moved upward in the drawing of FIG. 9 toward the composite boundary 34 between the front camera 7a and the left camera 7d, the three-dimensional object 41 disappears by more than half (25 cm or more) in the height direction.
- the lowest position of the three-dimensional object 41 in FIG. 9 is the constituent element of the boundary line 54b.
- a set of positions at the lowest position of the three-dimensional object 41 in FIG. 9 when the three-dimensional object 41 is gradually moved to the left in FIG. 9 and the same processing is repeated indicates a boundary line 54b.
- the boundary line 54c shown in the upper right part of FIG. 9 shows the boundary on the right side of FIG. That is, the boundary line 54 c indicates the outer edge of the host vehicle image 51.
- the boundary line 54d shown in the lower right part of FIG. 9 shows the upper boundary of the blind spot area 54 in FIG.
- the boundary line 54d indicates a boundary where a blind spot occurs when the three-dimensional object 41 is present in the region 22.
- the three-dimensional object 41 is moved downward in FIG. 9 toward the composite boundary 34 between the front camera 7a and the left camera 7d, the three-dimensional object 41 disappears by more than half (15 cm or more) in the width direction.
- the uppermost position in FIG. 9 of the three-dimensional object 41 is a constituent element of the boundary line 54d.
- a set of positions of the three-dimensional object 41 at the uppermost position on the paper surface of FIG. 9 when the three-dimensional object 41 is gradually moved to the left of the paper surface of FIG. 9 and the same processing is repeated indicates a boundary line 54d.
- the detection range of the left front corner sonar 12a is included within the angle of view of the second imaging means 14, and when the left front corner sonar 12a detects a three-dimensional object, the image processing means 3 performs the second imaging.
- the overhead image 40b is created by combining the captured image of the means 14 and the captured images of the four cameras 7a to 7d of the first imaging means, and the overhead image based on the captured image of the second imaging means 14 in the overhead image 40b.
- the detection range of the left front corner sonar 12a is included in this area. That is, the image processing means 3 moves the blind spot of the three-dimensional object outside the detection range of the left front corner sonar 12a.
- the three-dimensional object nearest to the host vehicle is not intended for all four corners of the host vehicle that the three-dimensional object nearest to the host vehicle disappears on the overhead view image, but on the left front that is likely to be a blind spot for the driver when using the right handle. Only targeted.
- the bird's-eye view image and the normal image are displayed in parallel to present a plurality of determination criteria to the driver, thereby improving the visibility of the three-dimensional object. Although improved, it is sufficient that at least an overhead image is displayed. That is, at least the processing of step S64 and step S65 in FIG. 6 may be performed. Specifically, the image processing unit 3 combines the captured image by the first imaging unit 7 and the captured image by the second imaging unit 14 into a bird's-eye view image based on common mapping data. And the display means 3 displays this bird's-eye view image.
- the continuity of the display image is maintained before and after the composition boundary of the overhead view image, and therefore, after detection of the three-dimensional object, generation of a blind spot and disappearance of the three-dimensional object at the composition boundary of the overhead image are suppressed while suppressing the driver's uncomfortable feeling. be able to.
- the second imaging means 14 provided at the left front corner and the left front corner sonar 12a are used, and the vicinity of the left front corner is likely to become a driver's blind spot.
- the three-dimensional object is prevented from disappearing on the bird's-eye view image, when the vehicle is the left steering wheel, the target is not the left front but the right front. That is, when the vehicle is a left steering wheel, the installation position of the second imaging means 14 in this embodiment is replaced with the right front corner sonar, and the left front corner sonar 12a is replaced with the right front corner sonar 12b.
- the second imaging means 14 images the front corner in the direction opposite to the handle position of the host vehicle among the four corners of the host vehicle.
- a detection range by the sonar 12 that detects a three-dimensional object existing in a front corner in a direction opposite to the steering wheel position of the host vehicle is included in the angle of view of the second imaging unit 14.
- the image processing unit 3 combines the captured image of the first imaging unit 7 and the captured image of the second imaging unit 14 to create an overhead image, and among the overhead images,
- the detection range by the sonar 12 may be included in the region of the bird's-eye view image based on the image captured by the two imaging means 14.
- the image processing unit 3 synthesizes the captured images of the first imaging unit 7 to create an overhead image. To do.
- the sonar for detecting the three-dimensional object before and after the vehicle in the sonar 12 is composed of the four sonars 12e to 12f. However, at least two sonars are used for detecting the three-dimensional object before and after the vehicle. What is necessary is just to comprise.
- the sonar 12 is used as the three-dimensional object detection means for detecting the three-dimensional object, but other means such as an infrared sensor may be used as long as the sensor detects the three-dimensional object.
- Embodiment 2 Next, the driving assistance apparatus of Embodiment 2 of this invention is demonstrated, referring drawings. Descriptions similar to those in the first embodiment are given the same reference numerals, and detailed descriptions thereof are omitted.
- FIG. 10 is a block diagram showing the configuration of the driving support apparatus according to the second embodiment of the present invention.
- the driving support device 1 further includes a selector 15 with respect to the driving support device of the first embodiment.
- the driving support device 1 is connected to the second imaging unit 14 via the selector 15.
- the selector 15 may also be included in the driving support device 1.
- the volatile memory 2 is connected to the second imaging unit 14 via the selector 15.
- the control means 5 controls the selector 15 based on the input signal from the sonar 12 and selects the second imaging means 14.
- the second imaging means 14 has four cameras each.
- the first imaging unit 7 and the second imaging unit 14 input captured images captured at predetermined time intervals to the volatile memory 2 of the driving support device 1.
- the 2nd imaging means 14 is attached to the vehicle body four corners of the own vehicle. The mounting position of the second imaging means 14 on the vehicle body will be described.
- FIG. 11 is a diagram for explaining the attachment positions of the first imaging means 7 and the second imaging means 14 to the vehicle.
- the second imaging means 14 includes a left front corner camera 14a, a right front corner camera 14b, a left rear corner camera 14c, and a right rear corner camera 14d.
- the selector 15 selects one camera from the left front corner camera 14a, the right front corner camera 14b, the left rear corner camera 14c, and the right rear corner camera 14d based on a command from the control means 5.
- the captured image of the selected camera is input to the volatile memory 2 of the driving support device 1 every predetermined time.
- the sonar 12 is composed of four sonars and is attached to the four corners of the vehicle body. The attachment position of the sonar 12 to the vehicle body will be described.
- FIG. 12 is a diagram for explaining the mounting position of the sonar 12 on the vehicle.
- the sonar 12 includes a left front corner sonar 12a, a right front corner sonar 12b, a left rear corner sonar 12c, and a right rear corner sonar 12d.
- the horizontal detection ranges 16a to 16d of the left front corner sonar 12a, right front corner sonar 12b, left rear corner sonar 12c, and right rear corner sonar 12d are set to 180 degrees or less.
- FIG. 13 is a diagram for explaining the angle of view of the left front corner camera 14a and the detection range of the horizontal plane of the left front corner sonar 12a with images.
- the relationship between the detection range of the sonar 12 and the angle of view of the second imaging unit 14 is the same as the relationship shown in FIG. 4 of the first embodiment. That is, the angle of view 17a of the left front corner camera 14a is set to approximately 180 degrees on a horizontal plane.
- the detection range 16a of the left front corner sonar 12a is included in the angle of view 17a of the left front corner camera 14a. That is, the entire detection range 16a of the left front corner sonar 12a is included in the angle of view 17a of the left front corner camera 14a.
- the left front corner camera 14a is mounted above the vehicle body relative to the left front corner sonar 12a. Further, it is desirable that the optical axes of the left front corner camera 14a and the left front corner sonar 12a are substantially equal.
- the relationship between the angle of view of the sonar 12 and the angle of view of the second imaging unit 14 is described using the angle of view of the left front corner camera 14 a and the detection range of the left front corner sonar 12 a.
- the same relationship holds for the other four corners of the vehicle. That is, the detection ranges of the right front corner sonar 12b, the left rear corner sonar 12c, and the right rear corner sonar 12d are included in the angles of view of the right front corner camera 14b, the left rear corner camera 14c, and the right rear corner camera 14d, respectively.
- FIG. 14 is a flowchart of the driving support process by the control means 5.
- Steps S71 to S73 are the same as steps S61 to S63 corresponding to FIG. 6 of the first embodiment.
- the control means 5 determines from the detection result of the sonar 12 whether or not a three-dimensional object exists within a predetermined range around the host vehicle. If NO in step S74, the process of step S71 is performed again. On the other hand, in the case of YES in step S74, as shown in step S75, the control means 5 specifies a three-dimensional object reaction location that exists at the shortest distance of the host vehicle.
- the control means 5 determines the corner where the sonar that detected the three-dimensional object is arranged as the three-dimensional object reaction point.
- the control means 5 is based on the distance signal between the own vehicle and the three-dimensional object input from the sonar 12. Then, the corner where the sonar that detects the three-dimensional object located closest to the host vehicle is determined as the three-dimensional object reaction point.
- step S76 the image processing unit 3 newly uses the captured image in the vicinity of the three-dimensional object reaction point at the shortest distance from the own vehicle specified in step S75 in addition to the captured image of the first imaging unit 7.
- a bird's-eye view image is created and the created bird's-eye view image is displayed on the display means 13. That is, the image processing means 3 selects the three-dimensional object reaction location of the sonar 12 identified in step S75 among the images taken by the four cameras 7a to 7d of the first imaging means 7 and the cameras 14a to 14d of the second imaging means 14.
- a bird's-eye view image is created using a captured image of the camera corresponding to the above.
- Differences between the overhead view image created in step S76 and the normal overhead view images created by the captured images of the four cameras 7a to 7d of the first imaging means 7 are the same as the differences described in the first embodiment. Therefore, detailed description is omitted.
- step S77 the control means 5 determines whether or not there is a three-dimensional object at the shortest distance from the vehicle in the traveling direction of the host vehicle.
- the control means 5 specifies the traveling direction of the host vehicle based on the gear signal input from the gear 11. That is, the control means 5 specifies from the gear signal that the traveling direction is forward when the shift lever is front and the traveling direction is backward when the shift lever is reverse. And the control means 5 compares this identified advancing direction with the three-dimensional object reaction location specified by step S75, and performs determination of step S77.
- control means 5 determines from the gear signal whether the three-dimensional object reaction point specified in step S75 is left front or right front.
- the control means 5 determines whether or not the three-dimensional object reaction location specified in step S75 is the left rear or the right rear.
- step S78 the control means 5 replaces the rear image displayed in parallel on the display means 13 in step S72 or the front image displayed in parallel on the display means 13 in step S73.
- the display unit 13 displays a captured image of a camera that captures the vicinity of the three-dimensional object reaction location at the shortest distance from the host vehicle identified in step S65 in the second imaging unit 14.
- the detection ranges of the sonars 12a to 12d are included in the angles of view of the corner cameras 14a to 14d, respectively, and when the sonars 12a to 12d detect the three-dimensional object, the three-dimensional object closest to the vehicle is detected.
- the corner camera corresponding to the sonar is selected by the selector 15.
- the image processing means 3 combines the picked-up images of the selected corner camera and the picked-up images of the four cameras a to 7d of the first picking-up means 7 to create an overhead image 40b.
- the sonar detection range is included in the region of the overhead image based on the image captured by the corner camera.
- the image processing means 3 moves the blind spot of the three-dimensional object outside the detection range of the sonar 12. Therefore, it is possible to prevent the three-dimensional object closest to the host vehicle from disappearing on the overhead view image of the display unit 13 even though the sonar 12 detects the three-dimensional object.
- the driving support device of the present invention makes it easy for the driver to recognize the three-dimensional object by displaying the three-dimensional object on the overhead view image without losing the three-dimensional object particularly when detecting the three-dimensional object at the four corners of the host vehicle using sonar or the like. Useful in terms.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Mechanical Engineering (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Closed-Circuit Television Systems (AREA)
- Traffic Control Systems (AREA)
- Image Processing (AREA)
Abstract
Description
以下、本発明の実施形態1の運転支援装置について図面を参照しながら説明する。なお、本実施形態において、車両は右ハンドルを一例として説明する。左ハンドルの場合は左右が逆になる。
次に、本発明の実施形態2の運転支援装置について図面を参照しながら説明する。実施形態1と同様の記載については同じ符号を付して詳細な説明を省略する。
3 画像処理手段
5 制御手段
7 第1撮像手段
12 ソナー
14 第2撮像手段
Claims (6)
- 自車両周囲を撮像する第1撮像手段と、
自車両の四隅のうち少なくとも自車両のハンドル位置と逆方向の前方コーナに存在する立体物を検出するセンサと、
自車両の四隅のうち少なくとも自車両のハンドル位置と逆方向の前方コーナを撮像する第2撮像手段と、
前記第1撮像手段から入力された自車両周囲の撮像画像を合成して俯瞰画像を作成する画像処理手段と、
前記画像処理手段が合成した俯瞰画像を表示する表示手段とを備え、
前記センサによる検出範囲は前記第2撮像手段の画角内に包含され、
前記センサが立体物を検出したとき、前記画像処理手段は、前記第2撮像手段の撮像画像と前記第1撮像手段の撮像画像とを合成して俯瞰画像を作成し、この俯瞰画像のうち前記第2撮像手段の撮像画像に基づく俯瞰画像の領域内に前記センサによる検出範囲を包含させる、
運転支援装置。 - 前記第1撮像手段は、自車両の前後左右をそれぞれ撮像する4つのカメラで構成され、前記センサは、自車両の四隅の立体物を検出する4つのセンサと自車両の前方と後方の立体物を検出する少なくとも2つのセンサで構成され、
前記画像処理手段は、前記センサのうち自車両のハンドル位置と逆方向の前方コーナに配置されたセンサが立体物を検出したときのみ、前記第2撮像手段によって撮像された撮像画像と前記第1撮像手段によって撮像された撮像画像を合成して俯瞰画像を作成し、前記センサのうち自車両のハンドル位置と逆方向の前方コーナに配置されたセンサが自立体物を検出しなかったとき、前記第1撮像手段によって撮像された撮像画像を合成した俯瞰画像を作成する、
請求項1に記載の運転支援装置。 - 前記センサは、自車両の四隅の立体物を検出する4つのセンサで構成され、
前記センサの検出結果に基づいて前記第2撮像手段を選択する選択手段をさらに備え、
前記センサが立体物を検出したとき、前記画像処理手段は、前記選択手段によって選択された前記第2撮像手段の撮像画像と前記第1撮像手段の撮像画像とを合成して俯瞰画像を作成し、この俯瞰画像のうち前記第2撮像手段の撮像画像に基づく俯瞰画像の領域内に前記センサの検出範囲を包含させる、
請求項1に記載の運転支援装置。 - 前記第1撮像手段は、自車両の前後左右をそれぞれ撮像する4つのカメラで構成され、前記センサは、自車両の四隅の立体物を検出する4つのセンサで構成され、前記第2撮像手段は、自車両の四隅をそれぞれ撮像する4つのカメラで構成され、
前記センサが立体物を検出したとき、前記選択手段はこの立体物を検出したセンサと同じ隅に配置されたカメラを選択し、
前記画像処理手段はこの選択されたカメラ画像と前記第1撮像手段によって撮像された4つのカメラ画像を合成して俯瞰画像を作成する、
請求項3に記載の運転支援装置。 - 前記センサと前記第2撮像手段は、光軸が略一致する、
請求項1に記載の運転支援装置。 - 前記俯瞰画像のうち前記第1撮像手段による撮像画像と前記第2撮像手段による撮像画像との合成境界は、前記第2撮像手段の画角に対して数度から十度内側に設定される、
請求項1に記載の運転支援装置。
Priority Applications (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2013531065A JP5979505B2 (ja) | 2011-08-26 | 2012-08-24 | 運転支援装置 |
US14/241,735 US10235575B2 (en) | 2011-08-26 | 2012-08-24 | Driving assistance apparatus |
EP12827913.0A EP2750382B1 (en) | 2011-08-26 | 2012-08-24 | Driving assistance apparatus |
CN201280040835.9A CN103782591B (zh) | 2011-08-26 | 2012-08-24 | 驾驶辅助装置 |
US16/263,159 US10970562B2 (en) | 2011-08-26 | 2019-01-31 | Display apparatus |
US17/194,595 US11657618B2 (en) | 2011-08-26 | 2021-03-08 | Display apparatus |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2011-184419 | 2011-08-26 | ||
JP2011184419 | 2011-08-26 | ||
JP2011184416 | 2011-08-26 | ||
JP2011-184416 | 2011-08-26 |
Related Child Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/241,735 A-371-Of-International US10235575B2 (en) | 2011-08-26 | 2012-08-24 | Driving assistance apparatus |
US16/263,159 Continuation US10970562B2 (en) | 2011-08-26 | 2019-01-31 | Display apparatus |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2013031161A1 true WO2013031161A1 (ja) | 2013-03-07 |
Family
ID=47755691
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2012/005321 WO2013031161A1 (ja) | 2011-08-26 | 2012-08-24 | 運転支援装置 |
Country Status (5)
Country | Link |
---|---|
US (3) | US10235575B2 (ja) |
EP (1) | EP2750382B1 (ja) |
JP (3) | JP5979505B2 (ja) |
CN (2) | CN107097721B (ja) |
WO (1) | WO2013031161A1 (ja) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2015230551A (ja) * | 2014-06-04 | 2015-12-21 | 日立建機株式会社 | 周囲物体検出システム及び運搬車両 |
WO2016031009A1 (ja) * | 2014-08-28 | 2016-03-03 | 国立大学法人東京大学 | 作業車両の表示システム、表示制御装置、作業車両、及び表示制御方法 |
CN108562914A (zh) * | 2018-05-29 | 2018-09-21 | 珠海市微半导体有限公司 | 一种基于盲区的红外检测装置和控制方法 |
Families Citing this family (26)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2013043661A1 (en) | 2011-09-21 | 2013-03-28 | Magna Electronics, Inc. | Vehicle vision system using image data transmission and power supply via a coaxial cable |
WO2013049597A1 (en) * | 2011-09-29 | 2013-04-04 | Allpoint Systems, Llc | Method and system for three dimensional mapping of an environment |
US10071687B2 (en) * | 2011-11-28 | 2018-09-11 | Magna Electronics Inc. | Vision system for vehicle |
US9956913B2 (en) * | 2013-03-28 | 2018-05-01 | Aisin Seiki Kabushiki Kaisha | Surroundings-monitoring device and computer program product |
US20150266420A1 (en) * | 2014-03-20 | 2015-09-24 | Honda Motor Co., Ltd. | Systems and methods for controlling a vehicle display |
DE102014107235A1 (de) * | 2014-05-22 | 2015-11-26 | Dr. Ing. H.C. F. Porsche Aktiengesellschaft | Verfahren zur Darstellung einer Fahrzeugumgebung auf einer Anzeigevorrichtung; eine Anzeigevorrichtung; ein System aus einer Mehrzahl an Bilderfassungseinheiten und einer Anzeigevorrichtung; ein Computerprogramm |
JP6224029B2 (ja) * | 2015-05-21 | 2017-11-01 | 富士通テン株式会社 | 画像処理装置および画像処理方法 |
JP6699427B2 (ja) * | 2015-11-17 | 2020-05-27 | 株式会社Jvcケンウッド | 車両用表示装置および車両用表示方法 |
KR101766077B1 (ko) * | 2015-12-08 | 2017-08-07 | 현대자동차주식회사 | 차량 주변 영상 정보 제공 시스템 및 방법 |
JP6621363B2 (ja) * | 2016-04-01 | 2019-12-18 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | 車両及び計測方法 |
US10594934B2 (en) | 2016-11-17 | 2020-03-17 | Bendix Commercial Vehicle Systems Llc | Vehicle display |
KR20180078970A (ko) * | 2016-12-30 | 2018-07-10 | 화남전자 주식회사 | 건설장비의 주변부 모니터링을 위한 카메라 배치 구조 |
KR102551099B1 (ko) * | 2017-01-13 | 2023-07-05 | 엘지이노텍 주식회사 | 어라운드뷰 제공장치, 제공방법 및 이를 구비한 차량 |
EP3537714B1 (en) * | 2017-02-28 | 2021-08-25 | JVC Kenwood Corporation | Bird's-eye-view video image generation device, bird's-eye-view video image generation system, bird's-eye-view video image generation method, and program |
CN110945537B (zh) * | 2017-07-28 | 2023-09-22 | 索尼互动娱乐股份有限公司 | 训练装置、识别装置、训练方法、识别方法和程序 |
EP3550830B1 (en) * | 2017-07-31 | 2021-09-08 | JVC Kenwood Corporation | Display control device, display control system, display control method, and program |
JP7201315B2 (ja) * | 2017-11-08 | 2023-01-10 | トヨタ自動車株式会社 | 電子ミラー装置 |
JP6809495B2 (ja) * | 2018-03-05 | 2021-01-06 | 株式会社デンソー | 画像作成装置 |
JP2019151304A (ja) * | 2018-03-06 | 2019-09-12 | アイシン精機株式会社 | 周辺監視装置 |
JP7135378B2 (ja) * | 2018-03-29 | 2022-09-13 | 株式会社アイシン | 周辺監視装置 |
DE102018214875A1 (de) * | 2018-08-31 | 2020-03-05 | Audi Ag | Verfahren und Anordnung zum Erzeugen einer Umgebungsrepräsentation eines Fahrzeugs und Fahrzeug mit einer solchen Anordnung |
KR20210147405A (ko) * | 2020-05-28 | 2021-12-07 | 삼성전자주식회사 | 객체 인식을 수행하는 전자 장치 및 이의 동작 방법 |
JP7424582B2 (ja) * | 2020-06-03 | 2024-01-30 | 株式会社ニフコ | 車載機器用ブラケット |
CN112009398A (zh) * | 2020-08-18 | 2020-12-01 | 三一专用汽车有限责任公司 | 车辆的控制方法、装置、车辆和存储介质 |
JP7400705B2 (ja) * | 2020-11-30 | 2023-12-19 | トヨタ自動車株式会社 | 撮像システム及び撮像装置 |
KR20220097656A (ko) * | 2020-12-30 | 2022-07-08 | 현대자동차주식회사 | 운전자 보조 장치, 차량 및 그 제어 방법 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2000064175A1 (fr) | 1999-04-16 | 2000-10-26 | Matsushita Electric Industrial Co., Ltd. | Dispositif de traitement d'images et systeme de surveillance |
JP2003169323A (ja) * | 2001-11-29 | 2003-06-13 | Clarion Co Ltd | 車両周囲監視装置 |
JP2004026012A (ja) * | 2002-06-25 | 2004-01-29 | Auto Network Gijutsu Kenkyusho:Kk | 車輌周辺視認装置 |
JP2006121587A (ja) | 2004-10-25 | 2006-05-11 | Nissan Motor Co Ltd | 運転支援装置及び俯瞰画像作成方法 |
JP2007104373A (ja) | 2005-10-05 | 2007-04-19 | Alpine Electronics Inc | 車両用画像表示装置 |
JP2008205914A (ja) * | 2007-02-21 | 2008-09-04 | Alpine Electronics Inc | 画像処理装置 |
Family Cites Families (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2005333565A (ja) | 2004-05-21 | 2005-12-02 | Auto Network Gijutsu Kenkyusho:Kk | 監視装置 |
JP4643995B2 (ja) | 2005-01-21 | 2011-03-02 | パナソニック株式会社 | 表示制御装置 |
US7782374B2 (en) * | 2005-03-03 | 2010-08-24 | Nissan Motor Co., Ltd. | Processor and processing method for generating a panoramic image for a vehicle |
JP4797441B2 (ja) * | 2005-05-20 | 2011-10-19 | トヨタ自動車株式会社 | 車両用画像処理装置 |
JP4956915B2 (ja) * | 2005-05-20 | 2012-06-20 | 日産自動車株式会社 | 映像表示装置及び映像表示方法 |
WO2007015446A1 (ja) | 2005-08-02 | 2007-02-08 | Nissan Motor Co., Ltd. | 車両周囲監視装置及び車両周囲監視方法 |
JP4807104B2 (ja) * | 2006-03-01 | 2011-11-02 | 日産自動車株式会社 | 車両周囲監視システム及び画像表示方法 |
JP2007235842A (ja) * | 2006-03-03 | 2007-09-13 | Denso Corp | 車載画像表示システム |
JP2008195268A (ja) | 2007-02-14 | 2008-08-28 | Denso Corp | 車両周辺監視装置 |
JP4985250B2 (ja) * | 2007-09-06 | 2012-07-25 | 株式会社デンソー | 駐車支援装置 |
JP5380941B2 (ja) | 2007-10-01 | 2014-01-08 | 日産自動車株式会社 | 駐車支援装置及び方法 |
JP2010250640A (ja) * | 2009-04-17 | 2010-11-04 | Sanyo Electric Co Ltd | 画像処理装置 |
DE102009036200A1 (de) * | 2009-08-05 | 2010-05-06 | Daimler Ag | Verfahren zur Überwachung einer Umgebung eines Fahrzeugs |
JP5503660B2 (ja) * | 2009-09-24 | 2014-05-28 | パナソニック株式会社 | 運転支援表示装置 |
JP2011077806A (ja) | 2009-09-30 | 2011-04-14 | Panasonic Corp | 車両周辺監視装置 |
-
2012
- 2012-08-24 CN CN201611201818.5A patent/CN107097721B/zh active Active
- 2012-08-24 CN CN201280040835.9A patent/CN103782591B/zh active Active
- 2012-08-24 JP JP2013531065A patent/JP5979505B2/ja active Active
- 2012-08-24 EP EP12827913.0A patent/EP2750382B1/en active Active
- 2012-08-24 US US14/241,735 patent/US10235575B2/en active Active
- 2012-08-24 WO PCT/JP2012/005321 patent/WO2013031161A1/ja active Application Filing
-
2016
- 2016-07-13 JP JP2016138376A patent/JP6241687B2/ja active Active
-
2017
- 2017-10-24 JP JP2017205423A patent/JP6388267B2/ja active Active
-
2019
- 2019-01-31 US US16/263,159 patent/US10970562B2/en active Active
-
2021
- 2021-03-08 US US17/194,595 patent/US11657618B2/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2000064175A1 (fr) | 1999-04-16 | 2000-10-26 | Matsushita Electric Industrial Co., Ltd. | Dispositif de traitement d'images et systeme de surveillance |
JP2003169323A (ja) * | 2001-11-29 | 2003-06-13 | Clarion Co Ltd | 車両周囲監視装置 |
JP2004026012A (ja) * | 2002-06-25 | 2004-01-29 | Auto Network Gijutsu Kenkyusho:Kk | 車輌周辺視認装置 |
JP2006121587A (ja) | 2004-10-25 | 2006-05-11 | Nissan Motor Co Ltd | 運転支援装置及び俯瞰画像作成方法 |
JP2007104373A (ja) | 2005-10-05 | 2007-04-19 | Alpine Electronics Inc | 車両用画像表示装置 |
JP2008205914A (ja) * | 2007-02-21 | 2008-09-04 | Alpine Electronics Inc | 画像処理装置 |
Non-Patent Citations (1)
Title |
---|
See also references of EP2750382A4 |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2015230551A (ja) * | 2014-06-04 | 2015-12-21 | 日立建機株式会社 | 周囲物体検出システム及び運搬車両 |
WO2016031009A1 (ja) * | 2014-08-28 | 2016-03-03 | 国立大学法人東京大学 | 作業車両の表示システム、表示制御装置、作業車両、及び表示制御方法 |
US10293752B2 (en) | 2014-08-28 | 2019-05-21 | The University Of Tokyo | Display system for work vehicle, display control device, work vehicle, and display control method |
CN108562914A (zh) * | 2018-05-29 | 2018-09-21 | 珠海市微半导体有限公司 | 一种基于盲区的红外检测装置和控制方法 |
Also Published As
Publication number | Publication date |
---|---|
JP5979505B2 (ja) | 2016-08-24 |
US20190163987A1 (en) | 2019-05-30 |
EP2750382A1 (en) | 2014-07-02 |
JP2016219025A (ja) | 2016-12-22 |
US11657618B2 (en) | 2023-05-23 |
US10235575B2 (en) | 2019-03-19 |
JPWO2013031161A1 (ja) | 2015-03-23 |
US10970562B2 (en) | 2021-04-06 |
JP6388267B2 (ja) | 2018-09-12 |
CN107097721B (zh) | 2020-03-06 |
EP2750382A4 (en) | 2014-08-13 |
US20210192230A1 (en) | 2021-06-24 |
CN107097721A (zh) | 2017-08-29 |
US20140218531A1 (en) | 2014-08-07 |
EP2750382B1 (en) | 2017-06-28 |
JP2018034794A (ja) | 2018-03-08 |
JP6241687B2 (ja) | 2017-12-06 |
CN103782591A (zh) | 2014-05-07 |
CN103782591B (zh) | 2017-02-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP6388267B2 (ja) | 運転支援装置 | |
JP4760831B2 (ja) | 車両周囲監視装置及び車両周囲監視方法 | |
JP5938635B2 (ja) | 運転支援装置 | |
JP3695319B2 (ja) | 車両周辺モニタ装置 | |
JP4639753B2 (ja) | 運転支援装置 | |
JP5321267B2 (ja) | 車両用画像表示装置及び俯瞰画像の表示方法 | |
JP6380410B2 (ja) | 画像表示制御装置および画像表示システム | |
JP4985250B2 (ja) | 駐車支援装置 | |
JP6730613B2 (ja) | 俯瞰映像生成装置、俯瞰映像生成システム、俯瞰映像生成方法およびプログラム | |
JP5251804B2 (ja) | 運転支援装置 | |
JP2018063294A (ja) | 表示制御装置 | |
JP2020068515A (ja) | 画像処理装置 | |
JP4927514B2 (ja) | 運転支援装置 | |
JP4595902B2 (ja) | 車両周辺画像表示システム及び車両周辺画像表示方法 | |
JP2004056219A (ja) | 車両周辺監視装置 | |
JP2012001126A (ja) | 車両用周辺監視装置 | |
KR20100021046A (ko) | 자동차용 표시장치 및 그 제어 방법 | |
JP2012166689A (ja) | 運転支援装置 | |
JP2007124097A (ja) | 車両周辺視認装置 | |
JP2012188058A (ja) | 表示装置 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 12827913 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2013531065 Country of ref document: JP Kind code of ref document: A |
|
REEP | Request for entry into the european phase |
Ref document number: 2012827913 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2012827913 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 14241735 Country of ref document: US |