[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

WO2018127962A1 - Reporting control device and reporting control method - Google Patents

Reporting control device and reporting control method Download PDF

Info

Publication number
WO2018127962A1
WO2018127962A1 PCT/JP2017/000193 JP2017000193W WO2018127962A1 WO 2018127962 A1 WO2018127962 A1 WO 2018127962A1 JP 2017000193 W JP2017000193 W JP 2017000193W WO 2018127962 A1 WO2018127962 A1 WO 2018127962A1
Authority
WO
WIPO (PCT)
Prior art keywords
event
unit
notification control
primary
recognition
Prior art date
Application number
PCT/JP2017/000193
Other languages
French (fr)
Japanese (ja)
Inventor
真規 塚田
下谷 光生
Original Assignee
三菱電機株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 三菱電機株式会社 filed Critical 三菱電機株式会社
Priority to JP2018560292A priority Critical patent/JP6661796B2/en
Priority to PCT/JP2017/000193 priority patent/WO2018127962A1/en
Publication of WO2018127962A1 publication Critical patent/WO2018127962A1/en

Links

Images

Classifications

    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/16Anti-collision systems

Definitions

  • the present invention relates to a notification control apparatus and a notification control method for notifying a driver of a situation around a host vehicle.
  • Patent Documents 1, 2, 3).
  • Patent Documents 1, 2, and 3 merely notify the driver of the actual situation, and do not notify the driver of a situation that may occur after the actual situation. Therefore, in Patent Documents 1, 2, and 3, it cannot be said that proper notification is given to the driver.
  • the present invention has been made to solve such a problem, and an object of the present invention is to provide a notification control device and a notification control method capable of performing appropriate notification to a driver.
  • a notification control apparatus includes a surrounding information acquisition unit that acquires at least surrounding information indicating a situation around the own vehicle, and the surrounding information acquired by the surrounding information acquisition unit.
  • a first event recognition unit for recognizing a first event which is a situation of an object or a person existing in the vicinity of the first event, and a situation that can occur after the first event recognized by the first event recognition unit,
  • a secondary event estimator that estimates a secondary event, which has a causal relationship with the situation of the primary event, according to a predetermined rule, and a secondary event estimated by the secondary event estimator is reported.
  • a notification control unit that performs control.
  • the notification control method acquires at least peripheral information indicating a situation around the host vehicle, and obtains a first event that is a situation of an object or person existing around the host vehicle from the acquired peripheral information. Recognizing and recognizing a secondary event that can occur after the recognized primary event and has a causal relationship with the situation of the primary event according to a predetermined rule Control to notify the secondary event is performed.
  • the notification control device includes at least a peripheral information acquisition unit that acquires peripheral information indicating a situation around the host vehicle, and an object existing around the host vehicle from the peripheral information acquired by the peripheral information acquisition unit.
  • a first event recognition unit that recognizes a first event that is a person's situation, and a situation that can occur after the first event recognized by the first event recognition unit, the situation and the causality of the first event
  • a secondary event estimator that estimates a secondary event that is a related situation according to a predetermined rule, and a notification controller that performs control to notify the secondary event estimated by the secondary event estimator Therefore, it is possible to appropriately notify the driver.
  • the notification control method acquires at least peripheral information indicating a situation around the host vehicle, and recognizes a primary event that is a situation of an object or a person existing around the host vehicle from the acquired surrounding information, A secondary event that is a situation that can occur after the recognized primary event and has a causal relationship with the situation of the primary event is estimated according to a predetermined rule. Therefore, it is possible to appropriately notify the driver.
  • FIG. 1 is a block diagram showing an example of the configuration of the notification control apparatus 1 according to the first embodiment. Note that FIG. 1 shows the minimum necessary components constituting the notification control apparatus according to the present embodiment.
  • the notification control device 1 includes a peripheral information acquisition unit 2, a primary event recognition unit 3, a secondary event estimation unit 4, and a notification control unit 5.
  • the peripheral information acquisition unit 2 acquires peripheral information indicating at least the situation around the host vehicle.
  • the primary event recognition unit 3 recognizes a primary event that is a situation of an object or person existing around the host vehicle from the peripheral information acquired by the peripheral information acquisition unit 2.
  • the primary event recognized by the primary event recognition unit may be a situation where the object or person is stationary, or a situation where the object or person is moving.
  • the secondary event estimation unit 4 is a situation that can occur after the primary event recognized by the primary event recognition unit 3 and has a causal relationship with the situation of the primary event. Is estimated according to a predetermined rule.
  • the notification control unit 5 performs control to notify the second event estimated by the second event estimation unit 4.
  • FIG. 2 is a block diagram showing an example of the configuration of the notification control device 6 according to another configuration.
  • the notification control device 6 includes a surrounding information acquisition unit 2, a primary event recognition unit 3, a secondary event estimation unit 4, a notification control unit 5, and a host vehicle information acquisition unit 7.
  • the peripheral information acquisition unit 2 is connected to the peripheral recognition device 10.
  • the notification control unit 5 is connected to a HUD (Head Up Display) 11.
  • the periphery recognition device 10 recognizes at least the situation around the host vehicle and outputs the recognized situation around the host vehicle to the periphery information acquisition unit 2 as the periphery information.
  • the periphery recognition device 10 is installed in the host vehicle, and is configured by, for example, a camera, a millimeter wave radar, or a combination thereof.
  • the periphery recognition device 10 is not limited to the camera and the millimeter wave radar, and may be any device that recognizes the surrounding situation of the host vehicle.
  • the peripheral information acquisition unit 2 acquires peripheral information output from the peripheral recognition device 10.
  • the own vehicle information acquisition unit 7 acquires own vehicle information, which is various information about the own vehicle, via an in-vehicle network provided in the own vehicle.
  • the host vehicle information acquisition unit 7 acquires the current position information of the host vehicle from the GPS (Global Positioning System) as host vehicle information.
  • GPS Global Positioning System
  • the map information acquisition unit 8 is constituted by a storage device such as a hard disk (HDD: Hard Disk Drive) or a semiconductor memory, for example, and acquires and stores map information.
  • the map information acquisition unit 8 may acquire map information from the outside.
  • the map information acquisition unit 8 may be acquired by downloading from an external server or the like via a communication network, or may be acquired by reading from a storage medium such as a memory.
  • the primary event recognition unit 3 recognizes a primary event that is a situation of an object or a person from an image that is peripheral information acquired by the peripheral information acquisition unit 2.
  • the first event recognition unit 3 may recognize the first event based on the current position information of the host vehicle acquired by the host vehicle information acquisition unit 7 and the map information acquired by the map information acquisition unit 8. Good.
  • the primary event recognition unit 3 may recognize the primary event by arbitrarily combining the peripheral information, the host vehicle information, and the map information.
  • the secondary event estimation unit 4 estimates the secondary event from the primary event recognized by the primary event recognition unit 3 according to a predetermined rule.
  • the predetermined rule is data that is converted into data and held by the secondary event estimation unit 4, and as an example, the primary event and the secondary event as shown in FIG. 3 are associated with each other.
  • the secondary event estimation unit 4 estimates “child jumping out” as the secondary event. That is, the secondary event estimation unit 4 estimates that when a ball pops out, a child pops out to take the ball.
  • the primary event recognition unit 3 can recognize the ball from an image captured by a camera that is the periphery recognition device 10, for example.
  • the primary event recognition unit 3 is not limited to recognizing the movement of the ball as the primary event, but may recognize that the ball is stationary as the primary event.
  • the secondary event estimating unit 4 performs “second jump of a person or motorcycle from behind the vehicle”. Estimated as the next event. That is, the secondary event estimation unit 4 estimates that a person or a motorcycle jumps out from behind the vehicle when there is a stopped vehicle.
  • the primary event recognition unit 3 can recognize a stopped vehicle from an image captured by a camera that is the periphery recognition device 10, for example.
  • the secondary event estimation unit 4 estimates “passing the train” as the secondary event. That is, the secondary event estimation unit 4 estimates that the train passes when there is a railroad crossing.
  • the primary event recognition unit 3 can recognize a railroad crossing from, for example, an image captured by a camera that is the periphery recognition device 10.
  • the primary event recognition unit 3 can recognize a railroad crossing based on the current position information of the host vehicle acquired by the host vehicle information acquisition unit 7 and the map information acquired by the map information acquisition unit 8.
  • the secondary event estimation unit 4 estimates “interrupt vehicle” as the secondary event. That is, when the vehicle lane on the road on which the host vehicle travels decreases, the secondary event estimation unit 4 estimates that the other vehicle will enter the vehicle lane on which the host vehicle is traveling.
  • the primary event recognition unit 3 can recognize that the vehicle lane of the road on which the host vehicle is traveling decreases from an image captured by a camera that is the periphery recognition device 10, for example.
  • the primary event recognition unit 3 is based on the current position information of the host vehicle acquired by the host vehicle information acquisition unit 7 and the map information acquired by the map information acquisition unit 8. It can be recognized that the lane decreases.
  • the first event when the second event is “interrupted vehicle” is not limited to the case where the vehicle lane decreases, but may be a vehicle lane in which the host vehicle is traveling, such as a junction of an expressway. A situation in which another vehicle is supposed to interrupt may be used as the first event.
  • the secondary event estimating unit 4 determines that “the object has fallen”. Is estimated as a secondary event.
  • the overhead road that intersects include a pedestrian bridge. That is, for example, when there is a person holding an object on a pedestrian bridge that intersects the road on which the host vehicle is traveling, the secondary event estimation unit 4 determines that the object that the person has is the road on which the host vehicle is traveling. Estimated to fall. In this case, the primary event recognizing unit 3 recognizes that there is a person who is facing the object on the intersecting overhead road from the image taken by the camera that is the periphery recognition device 10, for example. Can do.
  • the alarm image generation unit 9 generates a primary event alarm image indicating the primary event recognized by the primary event recognition unit 3. Further, the warning image generation unit 9 generates a secondary event warning image indicating the secondary event estimated by the secondary event estimation unit 4.
  • the notification control unit 5 performs control to display the first event alarm image or the second event alarm image generated by the alarm image generation unit 9 on the HUD 11.
  • a primary event warning image or a secondary event warning image is displayed on the HUD 11 using AR (Augmented Reality) technology. The driver can see the first event warning image or the second event warning image superimposed on the scenery seen through the windshield of the host vehicle.
  • the notification control unit 5 may perform control to simultaneously display the first event warning image and the second event warning image on the HUD 11.
  • FIG. 4 is a block diagram illustrating an example of a hardware configuration of the notification control device 6. The same applies to the notification control apparatus 1.
  • Peripheral information acquisition unit 2, primary event recognition unit 3, secondary event estimation unit 4, notification control unit 5, own vehicle information acquisition unit 7, map information acquisition unit 8, and alarm image generation unit in notification control device 6 Each function of 9 is realized by a processing circuit. That is, the notification control device 6 acquires peripheral information, recognizes a first event, estimates a second event, controls notification, acquires own vehicle information, acquires map information, and acquires first information.
  • a processing circuit is provided for generating a next event warning image and a second event warning image.
  • the processing circuit is a processor 12 (also called a central processing unit, processing unit, arithmetic unit, microprocessor, microcomputer, DSP (Digital Signal Processor)) that executes a program stored in the memory 13.
  • Peripheral information acquisition unit 2 primary event recognition unit 3, secondary event estimation unit 4, notification control unit 5, own vehicle information acquisition unit 7, map information acquisition unit 8, and alarm image generation unit in notification control device 6
  • Each function of 9 is realized by software, firmware, or a combination of software and firmware.
  • Software or firmware is described as a program and stored in the memory 13.
  • the processing circuit reads out and executes the program stored in the memory 13, thereby realizing the function of each unit. That is, the notification control device 6 obtains peripheral information, a step of recognizing a first event, a step of estimating a second event, a step of controlling notification, a step of acquiring own vehicle information, and map information.
  • a memory 13 is provided for storing a program in which the step of obtaining, the step of generating the primary event warning image and the secondary event warning image will be executed as a result.
  • these programs are the surrounding information acquisition part 2, the 1st event recognition part 3, the 2nd event estimation part 4, the alerting
  • the computer executes the procedure or method of the generation unit 9.
  • the memory is non-volatile or volatile such as RAM (Random Access Memory), ROM (Read Only Memory), flash memory, EPROM (Erasable Programmable Read Only Memory), EEPROM (Electrically Erasable Programmable Read Only Memory), etc. May be any other semiconductor memory, magnetic disk, flexible disk, optical disk, compact disk, mini disk, DVD (Digital Versatile Disc), or any other storage medium used in the future.
  • FIG. 5 is a flowchart showing an example of the operation of the notification control device 6.
  • the periphery recognition device 10 is a camera and photographs the front of the host vehicle.
  • step S11 the peripheral information acquisition unit 2 acquires an image captured by the peripheral recognition device 10 as peripheral information.
  • step S ⁇ b> 12 the primary event recognition unit 3 recognizes the primary event from the image acquired by the peripheral information acquisition unit 2.
  • the primary event recognition unit 3 recognizes the ball existing in front of the host vehicle as the primary event.
  • step S13 the alarm image generation unit 9 generates a primary event alarm image indicating the primary event recognized by the primary event recognition unit 3.
  • the notification control unit 5 performs control to display the first event warning image generated by the warning image generation unit 9 on the HUD 11.
  • FIG. 6 is a diagram illustrating an example when the first event warning image is displayed on the HUD 11.
  • the primary event warning image 14 is an image that emphasizes the ball itself.
  • the first event warning image 15 is an image that alerts the presence of an object that obstructs traveling ahead of the host vehicle, and an image that alerts the presence of a ball in the example of FIG. 6. It is.
  • the primary event warning image 14 and the primary event warning image 15 are not limited to being displayed together as shown in FIG. 6, and either one may be displayed.
  • step S14 the secondary event estimation unit 4 estimates the secondary event from the primary event recognized by the primary event recognition unit 3. Specifically, the secondary event estimation unit 4 estimates the secondary event based on data as shown in FIG. Here, the secondary event estimation unit 4 estimates “child's jumping out” as the secondary event.
  • step S15 the warning image generation unit 9 generates a secondary event warning image indicating the secondary event estimated by the secondary event estimation unit 4.
  • the notification control unit 5 performs control to display the second event warning image generated by the warning image generation unit 9 on the HUD 11.
  • FIG. 7 is a diagram showing an example when the second event warning image is displayed on the HUD 11.
  • the secondary event warning image 16 is an image in which a child jumps out in front of the host vehicle.
  • the secondary event warning image 16 is displayed at a predetermined position.
  • the position at which the second event warning image 16 is displayed may be arbitrarily set by the user.
  • the notification control devices 1 and 6 not only provide the primary event that is the actual situation, but also the secondary event that can occur after the primary event. Since the notification is made, it is possible to appropriately notify the driver. Therefore, the driver can take an appropriate response in consideration of the secondary event.
  • FIG. 8 is a block diagram showing an example of the configuration of the notification control device 17 according to the second embodiment.
  • the notification control device 17 is characterized by including a second event position estimation unit 18. Since other configurations and operations are the same as those in the first embodiment, detailed description thereof is omitted here.
  • the secondary event position estimation unit 18 estimates the occurrence position of the secondary event estimated by the secondary event estimation unit 4 based on the position where the primary event recognition unit 3 has recognized the primary event. .
  • the alarm image generation unit 9 generates a secondary event alarm image corresponding to the occurrence position estimated by the secondary event position estimation unit 18.
  • the function of the secondary event position estimation unit 18 in the notification control device 17 is realized by a processing circuit. That is, the secondary event position estimation unit 18 includes a processing circuit for estimating the occurrence position of the secondary event.
  • the processing circuit is, for example, a processor 12 that executes a program stored in the memory 13 as shown in FIG.
  • the function of the secondary event position estimation unit 18 in the notification control device 17 is realized by software, firmware, or a combination of software and firmware.
  • Software or firmware is described as a program and stored in the memory 13.
  • the processing circuit implements the function of the secondary event position estimating unit 18 by reading and executing the program stored in the memory 13. That is, the notification control device 17 includes a memory 13 for storing a program in which the step of estimating the occurrence position of the second event is executed as a result. It can also be said that this program causes the computer to execute the procedure or method of the secondary event position estimation unit 18.
  • FIG. 9 is a flowchart showing an example of the operation of the notification control device 17. Note that steps S21 to S24 in FIG. 9 correspond to steps S11 to S14 in FIG. Below, step S25 and step S26 are demonstrated.
  • the periphery recognition device 10 is a camera and shoots ahead of the host vehicle.
  • step S25 the secondary event position estimation unit 18 determines a predetermined position near the position where the primary event recognition unit 3 has recognized the primary event in the image acquired by the peripheral information acquisition unit 2. Presumed as the occurrence position of the secondary event.
  • step S26 the alarm image generation unit 9 generates a secondary event alarm image corresponding to the occurrence position of the secondary event estimated by the secondary event position estimation unit 18.
  • the notification control unit 5 performs control to display the second event warning image generated by the warning image generation unit 9 on the HUD 11.
  • FIG. 10 is a diagram illustrating an example when the second event warning image is displayed on the HUD 11.
  • the secondary event warning image 16 is displayed at the occurrence position estimated by the secondary event position estimation unit 18.
  • a state in which a child follows the ball and jumps out ahead of the host vehicle is expressed.
  • the notification control device 17 displays the second event at a position where the second event is highly likely to occur, so the driver performs the second event. It is possible to grasp more accurately than the first form.
  • FIG. 11 is a block diagram illustrating an example of the configuration of the notification control device 19 according to the third embodiment.
  • the notification control device 19 includes a recognition range restriction unit 20. Since other configurations and operations are the same as those in the second embodiment, detailed description thereof is omitted here.
  • the recognition range restriction unit 20 restricts the recognition range by the first event recognition unit 3 to a predetermined range including the occurrence position of the second event estimated by the second event position estimation unit 18.
  • the primary event recognition unit 3 recognizes the primary event within the range restricted by the recognition range restriction unit 20. For example, when the primary event recognition unit 3 recognizes the primary event from the image, the primary event recognition unit 3 recognizes the primary event within the range of the image restricted by the recognition range restriction unit 20. I do.
  • the function of the recognition range restriction unit 20 in the notification control device 19 is realized by a processing circuit. That is, the recognition range restriction unit 20 restricts the recognition range by the primary event recognition unit 3 to a predetermined range including the occurrence position of the secondary event estimated by the secondary event position estimation unit 18.
  • the processing circuit is provided.
  • the processing circuit is, for example, a processor 12 that executes a program stored in the memory 13 as shown in FIG.
  • the function of the recognition range restriction unit 20 in the notification control device 19 is realized by software, firmware, or a combination of software and firmware.
  • Software or firmware is described as a program and stored in the memory 13.
  • the processing circuit implements the function of the recognition range restriction unit 20 by reading and executing the program stored in the memory 13. That is, the notification control device 19 includes a step of limiting the recognition range by the primary event recognition unit 3 to a predetermined range including the occurrence position of the secondary event estimated by the secondary event position estimation unit 18.
  • a memory 13 is provided for storing a program to be executed as a result. It can also be said that this program causes a computer to execute the procedure or method of the recognition range restriction unit 20.
  • FIG. 12 is a flowchart showing an example of the operation of the notification control device 19. Note that steps S31 to S36 in FIG. 12 correspond to steps S21 to S26 in FIG. Below, step S37 and step S38 are demonstrated.
  • the periphery recognition device 10 is a camera and shoots ahead of the host vehicle.
  • step S37 the notification control device 19 determines whether or not to limit the recognition range by the primary event recognition unit 3.
  • the determination of whether or not to limit the recognition range by the primary event recognition unit 3 may be performed according to the resolution of the image acquired by the peripheral information acquisition unit 2. Specifically, when the resolution of the image is high, the processing load on the primary event recognition unit 3 increases, and therefore it may be determined that the recognition range by the primary event recognition unit 3 is limited. When the resolution of the image is low, it may be determined that the recognition range by the primary event recognition unit 3 is not limited because the processing load on the primary event recognition unit 3 is small.
  • the determination as to whether or not to limit the recognition range by the primary event recognition unit 3 may be arbitrarily set by the user. When limiting the recognition range by the primary event recognition part 3, it transfers to step S38. On the other hand, when the recognition range by the primary event recognition unit 3 is not limited, the process proceeds to step S31.
  • the recognition range restriction unit 20 restricts the recognition range by the primary event recognition unit 3 to a predetermined range including the occurrence position of the secondary event estimated by the secondary event position estimation unit 18. To do. Specifically, the recognition range restriction unit 20 restricts the recognition range by the primary event recognition unit 3 to, for example, a recognition range 21 as shown in FIG. Note that the recognition range 21 shown in FIG. 13 is shown for explanation, and is not actually displayed on the HUD 11.
  • the efficiency of recognizing the primary event by the primary event recognition unit 3 can be improved.
  • the secondary event can be displayed quickly.
  • the primary event recognition unit 3 can smoothly recognize the primary event. it can.
  • the primary event recognition unit 3 recognizes the primary event within the range limited by the recognition range limitation unit 20 for all images acquired by the peripheral information acquisition unit 2.
  • the present invention is not limited to this.
  • the primary event recognition unit 3 recognizes the primary event for all images within the range restricted by the recognition range restriction unit 20 and is outside the range restricted by the recognition range restriction unit 20. May recognize the primary event for the image acquired by the peripheral information acquisition unit 2 at predetermined intervals.
  • FIG. 14 is a block diagram showing an example of the configuration of the notification control device 22 according to the fourth embodiment.
  • the notification control device 22 is characterized by including a communication unit 23.
  • Other configurations and operations are the same as those in the third embodiment, and thus detailed description thereof is omitted here.
  • the communication unit 23 is communicably connected to each of the image recognition learning server 24 and the second event estimation data server 25 provided outside the notification control device 22, and the image recognition learning server 24 and the second event. Parameters and data to be described later are acquired from each of the estimated data servers 25.
  • the image recognition learning server 24 improves the recognition accuracy of the primary event by the primary event recognition unit 3 for parameters used when the primary event recognition unit 3 recognizes the primary event from the image.
  • Machine learning The secondary event estimation data server 25 accumulates data used when the secondary event estimation unit 4 estimates the secondary event.
  • the image recognition learning server 24 and the second event estimation data server 25 may be provided separately or may be provided integrally.
  • the function of the communication unit 23 in the notification control device 22 is realized by a processing circuit. That is, the communication unit 23 includes a processing circuit for communicating with each of the image recognition learning server 24 and the second event estimation data server 25.
  • the processing circuit is, for example, a processor 12 that executes a program stored in the memory 13 as shown in FIG.
  • the function of the communication unit 23 in the notification control device 22 is realized by software, firmware, or a combination of software and firmware.
  • Software or firmware is described as a program and stored in the memory 13.
  • the processing circuit implements the function of the communication unit 23 by reading and executing the program stored in the memory 13. That is, the notification control device 22 includes a memory 13 for storing a program in which a step of communicating with each of the image recognition learning server 24 and the second event estimation data server 25 is executed as a result. It can also be said that this program causes a computer to execute the procedure or method of the communication unit 23.
  • FIG. 15 is a flowchart showing an example of the operation of the notification control device 22. Note that steps S41 to S48 in FIG. 15 correspond to steps S31 to S38 in FIG. In step S47 of FIG. 15, if the recognition range by the primary event recognition unit 3 is not limited, the process proceeds to step S49. Hereinafter, step S49 will be described.
  • the periphery recognition device 10 is a camera and shoots ahead of the host vehicle.
  • step S49 the communication unit 23 acquires, from the image recognition learning server 24, parameters used when the primary event recognition unit 3 recognizes the primary event from the image.
  • the parameter used by the primary event recognition unit 3 is updated to the parameter acquired by the communication unit 23.
  • the communication unit 23 acquires data used when the secondary event estimation unit 4 estimates the secondary event from the secondary event estimation data server 25.
  • the data used by the second event estimation unit 4 is updated to the data acquired by the communication unit 23. Note that the process of step S49 is not limited to being performed after step S47 or step S48, and may be performed at an arbitrary timing.
  • FIG. 16 is a flowchart showing an example of operations of the image recognition learning server 24 and the secondary event estimation data server 25.
  • step S51 the image recognition learning server 24 recognizes the primary event recognition accuracy by the primary event recognition unit 3 for parameters used when the primary event recognition unit 3 recognizes the primary event from the image. Do machine learning that will improve. Machine-learned parameters are updated.
  • the secondary event estimation data server 25 accumulates data used when the secondary event estimation unit 4 estimates the secondary event. Specifically, the secondary event estimation data server 25 accumulates data associating the primary event and the secondary event as shown in FIG. 3, for example, and updates it as necessary.
  • step S ⁇ b> 53 the image recognition learning server 24 transmits parameters used when the primary event recognition unit 3 recognizes the primary event from the image to the communication unit 23 of the notification control device 22. Further, the secondary event estimation data server 25 transmits data used when the secondary event estimation unit 4 estimates the secondary event to the communication unit 23 of the notification control device 22. Note that the image recognition learning server 24 and the second event estimation data server 25 may transmit parameters and data simultaneously or separately. Further, the timing at which the image recognition learning server 24 and the second event estimation data server 25 transmit parameters and data may be a predetermined timing or an arbitrary timing set by the user.
  • the primary event recognition unit 3 updates the parameters used when the primary event recognition unit 3 recognizes the primary event from the image, so that the primary event recognition unit 3 It becomes possible to improve the recognition accuracy of the primary event. Moreover, the accuracy which the secondary event estimation part 4 estimates a secondary event can be improved by updating the data used when the secondary event estimation part 4 estimates a secondary event. .
  • FIG. 14 demonstrated the case where the notification control apparatus 19 shown in FIG. 11 was further equipped with the communication part 23, it is not restricted to this.
  • FIG. 17 is a block diagram showing an example of the configuration of the notification control device 26 according to the fifth embodiment.
  • the notification control device 26 includes a vehicle control unit 27. Since other configurations and operations are the same as those in the first embodiment, detailed description thereof is omitted here.
  • the vehicle control unit 27 controls the driving operation of the host vehicle based on the secondary event estimation result by the secondary event estimation unit 4. Specifically, the vehicle control unit 27 controls each operation of the accelerator, the brake, and the handle of the host vehicle.
  • the function of the vehicle control unit 27 in the notification control device 26 is realized by a processing circuit. That is, the vehicle control unit 27 includes a processing circuit for controlling the driving of the host vehicle based on the secondary event estimation result by the secondary event estimation unit 4.
  • the processing circuit is, for example, a processor 12 that executes a program stored in the memory 13 as shown in FIG.
  • the function of the vehicle control unit 27 in the notification control device 26 is realized by software, firmware, or a combination of software and firmware.
  • Software or firmware is described as a program and stored in the memory 13.
  • the processing circuit implements the function of the vehicle control unit 27 by reading and executing the program stored in the memory 13. That is, the vehicle control unit 27 stores a program in which the step of controlling the driving of the host vehicle is executed as a result based on the estimation result of the secondary event by the secondary event estimation unit 4.
  • a memory 13 is provided. It can also be said that this program causes a computer to execute the procedure or method of the vehicle control unit 27.
  • FIG. 18 is a flowchart showing an example of the operation of the notification control device 26. Note that steps S61 to S65 in FIG. 18 correspond to steps S11 to S15 in FIG. Hereinafter, step S66 will be described.
  • the periphery recognition device 10 is a camera and shoots ahead of the host vehicle.
  • step S66 the vehicle control unit 27 controls the driving operation of the host vehicle based on the secondary event estimation result by the secondary event estimation unit 4. Specifically, the vehicle control unit 27 reduces the reaction force of the steering wheel in the direction of avoiding the secondary event, operates the steering wheel in the direction of avoiding the secondary event, applies the brake early, etc. Control the driving operation of the vehicle. For example, referring to FIG. 3, when the second event is “interrupt vehicle”, vehicle control unit 27 performs control to decelerate the host vehicle.
  • the vehicle control unit 27 may control the driving operation of the host vehicle based on the recognition result by the primary event recognition unit 3. For example, referring to FIG. 3, when the first event is “ball jumping out”, the vehicle control unit 27 may perform control so that the steering wheel operation avoids the ball.
  • the notification control device 26 can not only notify the second event but also support the driving operation of the host vehicle.
  • FIG. 17 demonstrated the case where the notification control apparatus 6 shown in FIG. 2 was further provided with the vehicle control part 27, it is not restricted to this.
  • the vehicle control unit 27 may be provided in each of the notification control device 17 shown in FIG. 8, the notification control device 19 shown in FIG. 11, and the notification control device 22 shown in FIG.
  • the primary event recognition unit 3 recognizes the primary event has been described, but the present invention is not limited to this.
  • the first event recognized by another vehicle may be acquired from the other vehicle by inter-vehicle communication.
  • you may acquire the 1st event recognized by infrastructure facilities, such as a camera installed in the road, from the said infrastructure facility by road-to-vehicle communication.
  • the host vehicle can estimate the second event using the first event regarding the place that cannot be captured by the periphery recognition device 10 provided in the host vehicle.
  • the secondary event estimated by the secondary event estimation unit 4 may be transmitted to another vehicle.
  • Embodiments 1 to 5 the case where the camera that is the periphery recognition device 10 captures the front of the host vehicle has been described, but the present invention is not limited to this.
  • the camera that is the periphery recognition device 10 may capture all directions around the host vehicle or any direction.
  • the notification control device described above is not only a vehicle navigation device, that is, a car navigation device, but also a navigation device or navigation constructed as a system by appropriately combining a PND (Portable Navigation Device) and a server that can be mounted on a vehicle.
  • the present invention can be applied to apparatuses other than the apparatus. In this case, each function or each component of the notification control device is distributed and arranged in each function that constructs the system.
  • the function of the notification control device can be arranged in the server.
  • the vehicle side includes a peripheral recognition device 10 and a HUD 11
  • the server 28 includes a peripheral information acquisition unit 2, a primary event recognition unit 3, a secondary event estimation unit 4, and a notification control unit 5.
  • a notification control system can be constructed.
  • the notification control device 17 shown in FIG. 8 the notification control device 19 shown in FIG. 11, the notification control device 22 shown in FIG. 14, and the notification control device 26 shown in FIG.
  • the function of the notification control device can be arranged in the server and the mobile communication terminal 30.
  • the vehicle is provided with the periphery recognition device 10 and the HUD 11, and the server 29 includes the periphery information acquisition unit 2, the primary event recognition unit 3, the secondary event estimation unit 4, and the own vehicle information acquisition.
  • the unit 7, the map information acquisition unit 8, and the alarm image generation unit 9, and the notification control unit 5 in the mobile communication terminal 30 a notification control system can be constructed.
  • the notification control device 17 shown in FIG. 8 the notification control device 19 shown in FIG. 11, the notification control device 22 shown in FIG. 14, and the notification control device 26 shown in FIG.
  • software for executing the operation in the above embodiment may be incorporated in, for example, a server or a mobile communication terminal.
  • the above-described notification control method acquires at least peripheral information indicating a situation around the host vehicle, and is a situation of an object or a person existing around the host vehicle from the acquired peripheral information. Recognizing a primary event and estimating a secondary event that can occur after the recognized primary event and has a causal relationship with the situation of the primary event according to a predetermined rule Then, control for notifying the estimated secondary event is performed.

Landscapes

  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Traffic Control Systems (AREA)
  • Control Of Driving Devices And Active Controlling Of Vehicle (AREA)

Abstract

The purpose of the present invention is to provide a reporting control device and a reporting control method that make it possible to perform proper reporting to a driver. A reporting control device according to the present invention is provided with: a peripheral information obtaining unit that obtains peripheral information indicating at least the peripheral situation of a local vehicle; a primary-event recognition unit that recognizes, from the peripheral information obtained by the peripheral information obtaining unit, a primary event, which is a situation regarding an object or a person present around the local vehicle; a secondary-event estimation unit that estimates, in accordance with a predetermined rule, a secondary event, which is a situation that can occur after the primary event recognized by the primary-event recognition unit and that has a causal relationship with the situation of the primary event; and a reporting control unit that performs control to report the secondary event estimated by the secondary-event estimation unit.

Description

報知制御装置および報知制御方法Notification control device and notification control method
 本発明は、自車両の周辺の状況をドライバに報知する報知制御装置および報知制御方法に関する。 The present invention relates to a notification control apparatus and a notification control method for notifying a driver of a situation around a host vehicle.
 従来、自車両の周辺に存在する物体または人物を検知し、検知した物体または人物を強調して表示することによって、自車両のドライバに注意を喚起する技術が開示されている(例えば、特許文献1,2,3参照)。 2. Description of the Related Art Conventionally, a technique for alerting a driver of an own vehicle by detecting an object or person existing around the own vehicle and highlighting the detected object or person has been disclosed (for example, Patent Documents). 1, 2, 3).
特開平11-115660号公報Japanese Patent Laid-Open No. 11-115660 特開2008-62762号公報JP 2008-62762 A 特開2008-310376号公報JP 2008-310376 A
 例えば、走行中の自車両の前方にボールが存在している場合、この後、ボールを取るために子どもが自車両の前方に飛び出して来ることが考えられる。このような実際の状況の後に生じ得る状況を、実際の状況に加えてドライバに報知すれば、ドライバは今後生じ得る状況を考慮した適切な対応をとることができる。 For example, when a ball is present in front of the traveling vehicle, it is conceivable that the child will jump out in front of the traveling vehicle to take the ball. If a situation that may occur after such an actual situation is notified to the driver in addition to the actual situation, the driver can take an appropriate action in consideration of the situation that may occur in the future.
 特許文献1,2,3では、実際の状況をドライバに報知しているだけであり、実際の状況の後に生じ得る状況をドライバに報知していない。従って、特許文献1,2,3では、ドライバに対して適切な報知を行っているとはいえない。 Patent Documents 1, 2, and 3 merely notify the driver of the actual situation, and do not notify the driver of a situation that may occur after the actual situation. Therefore, in Patent Documents 1, 2, and 3, it cannot be said that proper notification is given to the driver.
 本発明は、このような問題を解決するためになされたものであり、ドライバに対して適切な報知を行うことが可能な報知制御装置および報知制御方法を提供することを目的とする。 The present invention has been made to solve such a problem, and an object of the present invention is to provide a notification control device and a notification control method capable of performing appropriate notification to a driver.
 上記の課題を解決するために、本発明による報知制御装置は、少なくとも自車両の周辺の状況を示す周辺情報を取得する周辺情報取得部と、周辺情報取得部が取得した周辺情報から、自車両の周辺に存在する物体または人物の状況である第1次事象を認識する第1次事象認識部と、第1次事象認識部が認識した第1次事象の後に生じ得る状況であって、第1次事象の状況と因果関係を有する状況である第2次事象を、予め定められた規則により推定する第2次事象推定部と、第2次事象推定部が推定した第2次事象を報知する制御を行う報知制御部とを備える。 In order to solve the above-described problem, a notification control apparatus according to the present invention includes a surrounding information acquisition unit that acquires at least surrounding information indicating a situation around the own vehicle, and the surrounding information acquired by the surrounding information acquisition unit. A first event recognition unit for recognizing a first event which is a situation of an object or a person existing in the vicinity of the first event, and a situation that can occur after the first event recognized by the first event recognition unit, A secondary event estimator that estimates a secondary event, which has a causal relationship with the situation of the primary event, according to a predetermined rule, and a secondary event estimated by the secondary event estimator is reported. A notification control unit that performs control.
 また、本発明による報知制御方法は、少なくとも自車両の周辺の状況を示す周辺情報を取得し、取得した周辺情報から、自車両の周辺に存在する物体または人物の状況である第1次事象を認識し、認識した第1次事象の後に生じ得る状況であって、第1次事象の状況と因果関係を有する状況である第2次事象を、予め定められた規則により推定し、推定した第2次事象を報知する制御を行う。 In addition, the notification control method according to the present invention acquires at least peripheral information indicating a situation around the host vehicle, and obtains a first event that is a situation of an object or person existing around the host vehicle from the acquired peripheral information. Recognizing and recognizing a secondary event that can occur after the recognized primary event and has a causal relationship with the situation of the primary event according to a predetermined rule Control to notify the secondary event is performed.
 本発明によると、報知制御装置は、少なくとも自車両の周辺の状況を示す周辺情報を取得する周辺情報取得部と、周辺情報取得部が取得した周辺情報から、自車両の周辺に存在する物体または人物の状況である第1次事象を認識する第1次事象認識部と、第1次事象認識部が認識した第1次事象の後に生じ得る状況であって、第1次事象の状況と因果関係を有する状況である第2次事象を、予め定められた規則により推定する第2次事象推定部と、第2次事象推定部が推定した第2次事象を報知する制御を行う報知制御部とを備えるため、ドライバに対して適切な報知を行うことが可能となる。 According to the present invention, the notification control device includes at least a peripheral information acquisition unit that acquires peripheral information indicating a situation around the host vehicle, and an object existing around the host vehicle from the peripheral information acquired by the peripheral information acquisition unit. A first event recognition unit that recognizes a first event that is a person's situation, and a situation that can occur after the first event recognized by the first event recognition unit, the situation and the causality of the first event A secondary event estimator that estimates a secondary event that is a related situation according to a predetermined rule, and a notification controller that performs control to notify the secondary event estimated by the secondary event estimator Therefore, it is possible to appropriately notify the driver.
 また、報知制御方法は、少なくとも自車両の周辺の状況を示す周辺情報を取得し、取得した周辺情報から、自車両の周辺に存在する物体または人物の状況である第1次事象を認識し、認識した第1次事象の後に生じ得る状況であって、第1次事象の状況と因果関係を有する状況である第2次事象を、予め定められた規則により推定し、推定した第2次事象を報知する制御を行うため、ドライバに対して適切な報知を行うことが可能となる。 In addition, the notification control method acquires at least peripheral information indicating a situation around the host vehicle, and recognizes a primary event that is a situation of an object or a person existing around the host vehicle from the acquired surrounding information, A secondary event that is a situation that can occur after the recognized primary event and has a causal relationship with the situation of the primary event is estimated according to a predetermined rule. Therefore, it is possible to appropriately notify the driver.
 本発明の目的、特徴、態様、および利点は、以下の詳細な説明と添付図面とによって、より明白となる。 The objects, features, aspects and advantages of the present invention will become more apparent from the following detailed description and the accompanying drawings.
本発明の実施の形態1による報知制御装置の構成の一例を示すブロック図である。It is a block diagram which shows an example of a structure of the alerting | reporting control apparatus by Embodiment 1 of this invention. 本発明の実施の形態1による報知制御装置の構成の一例を示すブロック図である。It is a block diagram which shows an example of a structure of the alerting | reporting control apparatus by Embodiment 1 of this invention. 本発明の実施の形態1による第1次事象および第2次事象の一例を示す図である。It is a figure which shows an example of the 1st event and the 2nd event by Embodiment 1 of this invention. 本発明の実施の形態1による報知制御装置のハードウェア構成の一例を示すブロック図である。It is a block diagram which shows an example of the hardware constitutions of the alerting | reporting control apparatus by Embodiment 1 of this invention. 本発明の実施の形態1による報知制御装置の動作の一例を示すフローチャートである。It is a flowchart which shows an example of operation | movement of the alerting | reporting control apparatus by Embodiment 1 of this invention. 本発明の実施の形態1による報知制御装置における表示の一例を示す図である。It is a figure which shows an example of the display in the alerting | reporting control apparatus by Embodiment 1 of this invention. 本発明の実施の形態1による報知制御装置における表示の一例を示す図である。It is a figure which shows an example of the display in the alerting | reporting control apparatus by Embodiment 1 of this invention. 本発明の実施の形態2による報知制御装置の構成の一例を示すブロック図である。It is a block diagram which shows an example of a structure of the alerting | reporting control apparatus by Embodiment 2 of this invention. 本発明の実施の形態2による報知制御装置の動作の一例を示すフローチャートである。It is a flowchart which shows an example of operation | movement of the alerting | reporting control apparatus by Embodiment 2 of this invention. 本発明の実施の形態2による報知制御装置における表示の一例を示す図である。It is a figure which shows an example of the display in the alerting | reporting control apparatus by Embodiment 2 of this invention. 本発明の実施の形態3による報知制御装置の構成の一例を示すブロック図である。It is a block diagram which shows an example of a structure of the alerting | reporting control apparatus by Embodiment 3 of this invention. 本発明の実施の形態3による報知制御装置の動作の一例を示すフローチャートである。It is a flowchart which shows an example of operation | movement of the alerting | reporting control apparatus by Embodiment 3 of this invention. 本発明の実施の形態3による報知制御装置における認識範囲を説明するための図である。It is a figure for demonstrating the recognition range in the alerting control apparatus by Embodiment 3 of this invention. 本発明の実施の形態4による報知制御装置の構成の一例を示すブロック図である。It is a block diagram which shows an example of a structure of the alerting | reporting control apparatus by Embodiment 4 of this invention. 本発明の実施の形態4による報知制御装置の動作の一例を示すフローチャートである。It is a flowchart which shows an example of operation | movement of the alerting | reporting control apparatus by Embodiment 4 of this invention. 本発明の実施の形態4による外部サーバの動作の一例を示すフローチャートである。It is a flowchart which shows an example of operation | movement of the external server by Embodiment 4 of this invention. 本発明の実施の形態5による報知制御装置の構成の一例を示すブロック図である。It is a block diagram which shows an example of a structure of the alerting | reporting control apparatus by Embodiment 5 of this invention. 本発明の実施の形態5による報知制御装置の動作の一例を示すフローチャートである。It is a flowchart which shows an example of operation | movement of the alerting | reporting control apparatus by Embodiment 5 of this invention. 本発明の実施の形態による報知制御システムの構成の一例を示すブロック図である。It is a block diagram which shows an example of a structure of the alerting | reporting control system by embodiment of this invention. 本発明の実施の形態による報知制御システムの構成の一例を示すブロック図である。It is a block diagram which shows an example of a structure of the alerting | reporting control system by embodiment of this invention.
 本発明の実施の形態について、図面に基づいて以下に説明する。 Embodiments of the present invention will be described below based on the drawings.
 <実施の形態1>
 <構成>
 まず、本発明の実施の形態1による報知制御装置の構成について説明する。
<Embodiment 1>
<Configuration>
First, the configuration of the notification control apparatus according to Embodiment 1 of the present invention will be described.
 図1は、本実施の形態1による報知制御装置1の構成の一例を示すブロック図である。なお、図1では、本実施の形態による報知制御装置を構成する必要最小限の構成要素を示している。 FIG. 1 is a block diagram showing an example of the configuration of the notification control apparatus 1 according to the first embodiment. Note that FIG. 1 shows the minimum necessary components constituting the notification control apparatus according to the present embodiment.
 図1に示すように、報知制御装置1は、周辺情報取得部2と、第1次事象認識部3と、第2次事象推定部4と、報知制御部5とを備えている。周辺情報取得部2は、少なくとも自車両の周辺の状況を示す周辺情報を取得する。第1次事象認識部3は、周辺情報取得部2が取得した周辺情報から、自車両の周辺に存在する物体または人物の状況である第1次事象を認識する。第1次事象認識部が認識する第1次事象は、物体または人物が静止している状況であってもよく、物体または人物が動いている状況であってもよい。 As shown in FIG. 1, the notification control device 1 includes a peripheral information acquisition unit 2, a primary event recognition unit 3, a secondary event estimation unit 4, and a notification control unit 5. The peripheral information acquisition unit 2 acquires peripheral information indicating at least the situation around the host vehicle. The primary event recognition unit 3 recognizes a primary event that is a situation of an object or person existing around the host vehicle from the peripheral information acquired by the peripheral information acquisition unit 2. The primary event recognized by the primary event recognition unit may be a situation where the object or person is stationary, or a situation where the object or person is moving.
 第2次事象推定部4は、第1次事象認識部3が認識した第1次事象の後に生じ得る状況であって、第1次事象の状況と因果関係を有する状況である第2次事象を、予め定められた規則により推定する。報知制御部5は、第2次事象推定部4が推定した第2次事象を報知する制御を行う。 The secondary event estimation unit 4 is a situation that can occur after the primary event recognized by the primary event recognition unit 3 and has a causal relationship with the situation of the primary event. Is estimated according to a predetermined rule. The notification control unit 5 performs control to notify the second event estimated by the second event estimation unit 4.
 次に、図1に示す報知制御装置1を含む報知制御装置の他の構成について説明する。 Next, another configuration of the notification control device including the notification control device 1 shown in FIG. 1 will be described.
 図2は、他の構成に係る報知制御装置6の構成の一例を示すブロック図である。 FIG. 2 is a block diagram showing an example of the configuration of the notification control device 6 according to another configuration.
 図2に示すように、報知制御装置6は、周辺情報取得部2と、第1次事象認識部3と、第2次事象推定部4と、報知制御部5と、自車両情報取得部7と、地図情報取得部8と、警報画像生成部9とを備えている。周辺情報取得部2は、周辺認識装置10に接続されている。報知制御部5は、HUD(Head Up Display)11に接続されている。 As shown in FIG. 2, the notification control device 6 includes a surrounding information acquisition unit 2, a primary event recognition unit 3, a secondary event estimation unit 4, a notification control unit 5, and a host vehicle information acquisition unit 7. A map information acquisition unit 8 and an alarm image generation unit 9. The peripheral information acquisition unit 2 is connected to the peripheral recognition device 10. The notification control unit 5 is connected to a HUD (Head Up Display) 11.
 周辺認識装置10は、少なくとも自車両の周辺の状況を認識し、認識した自車両の周辺の状況を周辺情報として周辺情報取得部2に出力する。周辺認識装置10は、自車両に設置されており、例えばカメラ、ミリ波レーダ、またはこれらの組み合わせによって構成されている。なお、周辺認識装置10は、カメラおよびミリ波レーダに限らず、自車両の周辺状況を認識するものであればよい。周辺情報取得部2は、周辺認識装置10から出力された周辺情報を取得する。 The periphery recognition device 10 recognizes at least the situation around the host vehicle and outputs the recognized situation around the host vehicle to the periphery information acquisition unit 2 as the periphery information. The periphery recognition device 10 is installed in the host vehicle, and is configured by, for example, a camera, a millimeter wave radar, or a combination thereof. The periphery recognition device 10 is not limited to the camera and the millimeter wave radar, and may be any device that recognizes the surrounding situation of the host vehicle. The peripheral information acquisition unit 2 acquires peripheral information output from the peripheral recognition device 10.
 自車両情報取得部7は、自車両内に設けられた車内ネットワークを介して、自車両に関する種々の情報である自車両情報を取得する。例えば、自車両情報取得部7は、GPS(Global Positioning System)から自車両の現在の位置情報を自車両情報として取得する。 The own vehicle information acquisition unit 7 acquires own vehicle information, which is various information about the own vehicle, via an in-vehicle network provided in the own vehicle. For example, the host vehicle information acquisition unit 7 acquires the current position information of the host vehicle from the GPS (Global Positioning System) as host vehicle information.
 地図情報取得部8は、例えば、ハードディスク(HDD:Hard Disk Drive)または半導体メモリ等の記憶装置から構成されており、地図情報を取得して格納している。なお、地図情報取得部8は、外部から地図情報を取得してもよい。例えば、地図情報取得部8は、外部のサーバ等から通信ネットワークを介してダウンロードすることによって取得してもよく、メモリ等の記憶媒体から読み出すことによって取得してもよい。 The map information acquisition unit 8 is constituted by a storage device such as a hard disk (HDD: Hard Disk Drive) or a semiconductor memory, for example, and acquires and stores map information. The map information acquisition unit 8 may acquire map information from the outside. For example, the map information acquisition unit 8 may be acquired by downloading from an external server or the like via a communication network, or may be acquired by reading from a storage medium such as a memory.
 第1次事象認識部3は、例えば周辺認識装置10がカメラである場合、周辺情報取得部2が取得した周辺情報である画像から、物体または人物の状況である第1次事象を認識する。また、第1次事象認識部3は、自車両情報取得部7が取得した自車両の現在位置情報、および地図情報取得部8が取得した地図情報に基づいて第1次事象を認識してもよい。なお、第1次事象認識部3は、周辺情報、自車両情報、および地図情報を任意に組み合わせて第1次事象を認識してもよい。 For example, when the peripheral recognition device 10 is a camera, the primary event recognition unit 3 recognizes a primary event that is a situation of an object or a person from an image that is peripheral information acquired by the peripheral information acquisition unit 2. The first event recognition unit 3 may recognize the first event based on the current position information of the host vehicle acquired by the host vehicle information acquisition unit 7 and the map information acquired by the map information acquisition unit 8. Good. Note that the primary event recognition unit 3 may recognize the primary event by arbitrarily combining the peripheral information, the host vehicle information, and the map information.
 第2次事象推定部4は、第1次事象認識部3が認識した第1次事象から第2次事象を予め定められた規則により推定する。予め定められた規則は、データ化されて第2次事象推定部4が保持しており、一例として図3に示すような第1次事象と第2次事象とを対応付けたデータである。 The secondary event estimation unit 4 estimates the secondary event from the primary event recognized by the primary event recognition unit 3 according to a predetermined rule. The predetermined rule is data that is converted into data and held by the secondary event estimation unit 4, and as an example, the primary event and the secondary event as shown in FIG. 3 are associated with each other.
 例えば、第1次事象認識部3が「ボールの飛び出し」を第1次事象として認識した場合、第2次事象推定部4は「子どもの飛び出し」を第2次事象として推定する。すなわち、第2次事象推定部4は、ボールが飛び出して来た場合、ボールを取るために子どもが飛び出して来ると推定する。この場合、第1次事象認識部3は、例えば周辺認識装置10であるカメラが撮影した画像からボールを認識することができる。なお、第1次事象認識部3は、ボールの動きを第1次事象として認識することに限らず、ボールが静止していることを第1次事象として認識してもよい。 For example, when the primary event recognition unit 3 recognizes “ball jumping out” as the primary event, the secondary event estimation unit 4 estimates “child jumping out” as the secondary event. That is, the secondary event estimation unit 4 estimates that when a ball pops out, a child pops out to take the ball. In this case, the primary event recognition unit 3 can recognize the ball from an image captured by a camera that is the periphery recognition device 10, for example. The primary event recognition unit 3 is not limited to recognizing the movement of the ball as the primary event, but may recognize that the ball is stationary as the primary event.
 第1次事象認識部3が道路脇に停車中の「停車車両」を第1次事象として認識した場合、第2次事象推定部4は「車両の陰から人またはバイクの飛び出し」を第2次事象として推定する。すなわち、第2次事象推定部4は、停車車両が存在している場合、車両の陰から人またはバイクが飛び出して来ると推定する。この場合、第1次事象認識部3は、例えば周辺認識装置10であるカメラが撮影した画像から停車車両を認識することができる。 When the primary event recognizing unit 3 recognizes “stopped vehicle” parked on the side of the road as the first event, the secondary event estimating unit 4 performs “second jump of a person or motorcycle from behind the vehicle”. Estimated as the next event. That is, the secondary event estimation unit 4 estimates that a person or a motorcycle jumps out from behind the vehicle when there is a stopped vehicle. In this case, the primary event recognition unit 3 can recognize a stopped vehicle from an image captured by a camera that is the periphery recognition device 10, for example.
 第1次事象認識部3が「踏切」を第1次事象として認識した場合、第2次事象推定部4は「電車の通過」を第2次事象として推定する。すなわち、第2次事象推定部4は、踏切がある場合、電車が通過すると推定する。この場合、第1次事象認識部3は、例えば周辺認識装置10であるカメラが撮影した画像から踏切を認識することができる。あるいは、第1次事象認識部3は、自車両情報取得部7が取得した自車両の現在位置情報、および地図情報取得部8が取得した地図情報に基づいて踏切を認識することができる。 When the primary event recognition unit 3 recognizes “crossing” as the primary event, the secondary event estimation unit 4 estimates “passing the train” as the secondary event. That is, the secondary event estimation unit 4 estimates that the train passes when there is a railroad crossing. In this case, the primary event recognition unit 3 can recognize a railroad crossing from, for example, an image captured by a camera that is the periphery recognition device 10. Alternatively, the primary event recognition unit 3 can recognize a railroad crossing based on the current position information of the host vehicle acquired by the host vehicle information acquisition unit 7 and the map information acquired by the map information acquisition unit 8.
 第1次事象認識部3が「車両レーンの減少」を第1次事象として認識した場合、第2次事象推定部4は「割り込み車両」を第2次事象として推定する。すなわち、第2次事象推定部4は、自車両が走行する道路の車両レーンが減少する場合、自車両が走行中の車両レーンに他車両が割り込んでくると推定する。この場合、第1次事象認識部3は、例えば周辺認識装置10であるカメラが撮影した画像から自車両が走行する道路の車両レーンが減少することを認識することができる。あるいは、第1次事象認識部3は、自車両情報取得部7が取得した自車両の現在位置情報、および地図情報取得部8が取得した地図情報に基づいて、自車両が走行する道路の車両レーンが減少することを認識することができる。なお、自車両が走行する道路の車両レーンが減少する場合としては、例えば、車両レーンが2車線から1車線に減少する場合などが挙げられる。また、第2次事象が「割り込み車両」とされる場合の第1次事象としては、車両レーンが減少する場合に限らず、例えば高速道路の合流点など、自車両が走行中の車両レーンに他車両が割り込んでくることが想定される状況を第1次事象としてもよい。 When the primary event recognition unit 3 recognizes “vehicle lane reduction” as the primary event, the secondary event estimation unit 4 estimates “interrupt vehicle” as the secondary event. That is, when the vehicle lane on the road on which the host vehicle travels decreases, the secondary event estimation unit 4 estimates that the other vehicle will enter the vehicle lane on which the host vehicle is traveling. In this case, the primary event recognition unit 3 can recognize that the vehicle lane of the road on which the host vehicle is traveling decreases from an image captured by a camera that is the periphery recognition device 10, for example. Alternatively, the primary event recognition unit 3 is based on the current position information of the host vehicle acquired by the host vehicle information acquisition unit 7 and the map information acquired by the map information acquisition unit 8. It can be recognized that the lane decreases. In addition, as a case where the vehicle lane of the road where the own vehicle travels decreases, for example, the case where the vehicle lane decreases from two lanes to one lane can be cited. In addition, the first event when the second event is “interrupted vehicle” is not limited to the case where the vehicle lane decreases, but may be a vehicle lane in which the host vehicle is traveling, such as a junction of an expressway. A situation in which another vehicle is supposed to interrupt may be used as the first event.
 第1次事象認識部3が「交差する頭上道路上に、こちらを向いて物を持っている人物」を第1次事象として認識した場合、第2次事象推定部4は「物の落下」を第2次事象として推定する。交差する頭上道路とは、例えば歩道橋などが挙げられる。すなわち、第2次事象推定部4は、例えば自車両が走行中の道路を交差する歩道橋に物を持っている人物がいる場合、当該人物が持っている物が、自車両が走行中の道路に落下すると推定する。この場合、第1次事象認識部3は、例えば周辺認識装置10であるカメラが撮影した画像から、交差する頭上道路上に、こちらを向いて物を持っている人物がいることを認識することができる。 When the primary event recognizing unit 3 recognizes “a person holding an object facing this on an intersecting overhead road” as the primary event, the secondary event estimating unit 4 determines that “the object has fallen”. Is estimated as a secondary event. Examples of the overhead road that intersects include a pedestrian bridge. That is, for example, when there is a person holding an object on a pedestrian bridge that intersects the road on which the host vehicle is traveling, the secondary event estimation unit 4 determines that the object that the person has is the road on which the host vehicle is traveling. Estimated to fall. In this case, the primary event recognizing unit 3 recognizes that there is a person who is facing the object on the intersecting overhead road from the image taken by the camera that is the periphery recognition device 10, for example. Can do.
 警報画像生成部9は、第1次事象認識部3が認識した第1次事象を示す第1次事象警報画像を生成する。また、警報画像生成部9は、第2次事象推定部4が推定した第2次事象を示す第2次事象警報画像を生成する。 The alarm image generation unit 9 generates a primary event alarm image indicating the primary event recognized by the primary event recognition unit 3. Further, the warning image generation unit 9 generates a secondary event warning image indicating the secondary event estimated by the secondary event estimation unit 4.
 報知制御部5は、警報画像生成部9が生成した第1次事象警報画像または第2次事象警報画像をHUD11に表示する制御を行う。HUD11には、AR(Augmented Reality)技術を用いて第1次事象警報画像または第2次事象警報画像が表示される。ドライバは、自車両のフロントガラス越しに見える景色に、第1次事象警報画像または第2次事象警報画像が重畳して表示されているように見ることができる。なお、報知制御部5は、第1次事象警報画像および第2次事象警報画像を同時にHUD11に表示する制御を行ってもよい。 The notification control unit 5 performs control to display the first event alarm image or the second event alarm image generated by the alarm image generation unit 9 on the HUD 11. A primary event warning image or a secondary event warning image is displayed on the HUD 11 using AR (Augmented Reality) technology. The driver can see the first event warning image or the second event warning image superimposed on the scenery seen through the windshield of the host vehicle. Note that the notification control unit 5 may perform control to simultaneously display the first event warning image and the second event warning image on the HUD 11.
 図4は、報知制御装置6のハードウェア構成の一例を示すブロック図である。なお、報知制御装置1についても同様である。 FIG. 4 is a block diagram illustrating an example of a hardware configuration of the notification control device 6. The same applies to the notification control apparatus 1.
 報知制御装置6における周辺情報取得部2、第1次事象認識部3、第2次事象推定部4、報知制御部5、自車両情報取得部7、地図情報取得部8、および警報画像生成部9の各機能は、処理回路により実現される。すなわち、報知制御装置6は、周辺情報を取得し、第1次事象を認識し、第2次事象を推定し、報知を制御し、自車両情報を取得し、地図情報を取得し、第1次事象警報画像および第2次事象警報画像を生成するための処理回路を備える。処理回路は、メモリ13に格納されたプログラムを実行するプロセッサ12(中央処理装置、処理装置、演算装置、マイクロプロセッサ、マイクロコンピュータ、DSP(Digital Signal Processor)ともいう)である。 Peripheral information acquisition unit 2, primary event recognition unit 3, secondary event estimation unit 4, notification control unit 5, own vehicle information acquisition unit 7, map information acquisition unit 8, and alarm image generation unit in notification control device 6 Each function of 9 is realized by a processing circuit. That is, the notification control device 6 acquires peripheral information, recognizes a first event, estimates a second event, controls notification, acquires own vehicle information, acquires map information, and acquires first information. A processing circuit is provided for generating a next event warning image and a second event warning image. The processing circuit is a processor 12 (also called a central processing unit, processing unit, arithmetic unit, microprocessor, microcomputer, DSP (Digital Signal Processor)) that executes a program stored in the memory 13.
 報知制御装置6における周辺情報取得部2、第1次事象認識部3、第2次事象推定部4、報知制御部5、自車両情報取得部7、地図情報取得部8、および警報画像生成部9の各機能は、ソフトウェア、ファームウェア、またはソフトウェアとファームウェアとの組み合わせにより実現される。ソフトウェアまたはファームウェアは、プログラムとして記述され、メモリ13に格納される。処理回路は、メモリ13に記憶されたプログラムを読み出して実行することにより、各部の機能を実現する。すなわち、報知制御装置6は、周辺情報を取得するステップ、第1次事象を認識するステップ、第2次事象を推定するステップ、報知を制御するステップ、自車両情報を取得するステップ、地図情報を取得するステップ、第1次事象警報画像および第2次事象警報画像を生成するステップが結果的に実行されることになるプログラムを格納するためのメモリ13を備える。また、これらのプログラムは、周辺情報取得部2、第1次事象認識部3、第2次事象推定部4、報知制御部5、自車両情報取得部7、地図情報取得部8、および警報画像生成部9の手順または方法をコンピュータに実行させるものであるともいえる。ここで、メモリとは、例えば、RAM(Random Access Memory)、ROM(Read Only Memory)、フラッシュメモリ、EPROM(Erasable Programmable Read Only Memory)、EEPROM(Electrically Erasable Programmable Read Only Memory)等の不揮発性または揮発性の半導体メモリ、磁気ディスク、フレキシブルディスク、光ディスク、コンパクトディスク、ミニディスク、DVD(Digital Versatile Disc)等、または、今後使用されるあらゆる記憶媒体であってもよい。 Peripheral information acquisition unit 2, primary event recognition unit 3, secondary event estimation unit 4, notification control unit 5, own vehicle information acquisition unit 7, map information acquisition unit 8, and alarm image generation unit in notification control device 6 Each function of 9 is realized by software, firmware, or a combination of software and firmware. Software or firmware is described as a program and stored in the memory 13. The processing circuit reads out and executes the program stored in the memory 13, thereby realizing the function of each unit. That is, the notification control device 6 obtains peripheral information, a step of recognizing a first event, a step of estimating a second event, a step of controlling notification, a step of acquiring own vehicle information, and map information. A memory 13 is provided for storing a program in which the step of obtaining, the step of generating the primary event warning image and the secondary event warning image will be executed as a result. Moreover, these programs are the surrounding information acquisition part 2, the 1st event recognition part 3, the 2nd event estimation part 4, the alerting | reporting control part 5, the own vehicle information acquisition part 7, the map information acquisition part 8, and an alarm image. It can also be said that the computer executes the procedure or method of the generation unit 9. Here, the memory is non-volatile or volatile such as RAM (Random Access Memory), ROM (Read Only Memory), flash memory, EPROM (Erasable Programmable Read Only Memory), EEPROM (Electrically Erasable Programmable Read Only Memory), etc. May be any other semiconductor memory, magnetic disk, flexible disk, optical disk, compact disk, mini disk, DVD (Digital Versatile Disc), or any other storage medium used in the future.
 <動作>
 次に、報知制御装置6の動作について説明する。
<Operation>
Next, the operation of the notification control device 6 will be described.
 図5は、報知制御装置6の動作の一例を示すフローチャートである。なお、以下では、周辺認識装置10はカメラであり、自車両の前方を撮影するものとする。 FIG. 5 is a flowchart showing an example of the operation of the notification control device 6. In the following, it is assumed that the periphery recognition device 10 is a camera and photographs the front of the host vehicle.
 ステップS11において、周辺情報取得部2は、周辺認識装置10が撮影した画像を周辺情報として取得する。ステップS12において、第1次事象認識部3は、周辺情報取得部2が取得した画像から第1次事象を認識する。ここでは、第1次事象認識部3は、自車両の前方に存在するボールを第1次事象として認識する。 In step S11, the peripheral information acquisition unit 2 acquires an image captured by the peripheral recognition device 10 as peripheral information. In step S <b> 12, the primary event recognition unit 3 recognizes the primary event from the image acquired by the peripheral information acquisition unit 2. Here, the primary event recognition unit 3 recognizes the ball existing in front of the host vehicle as the primary event.
 ステップS13において、警報画像生成部9は、第1次事象認識部3が認識した第1次事象を示す第1次事象警報画像を生成する。報知制御部5は、警報画像生成部9が生成した第1次事象警報画像をHUD11に表示する制御を行う。 In step S13, the alarm image generation unit 9 generates a primary event alarm image indicating the primary event recognized by the primary event recognition unit 3. The notification control unit 5 performs control to display the first event warning image generated by the warning image generation unit 9 on the HUD 11.
 図6は、第1次事象警報画像をHUD11に表示した場合の一例を示す図である。図6において、第1次事象警報画像14は、ボール自体を強調するような画像である。第1次事象警報画像15は、自車両の前方に走行を妨げるような物体が存在していること注意喚起する画像であり、図6の例ではボールが存在していることを注意喚起する画像である。なお、第1次事象警報画像14および第1次事象警報画像15は、図6に示すように一緒に表示することに限らず、いずれか一方を表示してもよい。 FIG. 6 is a diagram illustrating an example when the first event warning image is displayed on the HUD 11. In FIG. 6, the primary event warning image 14 is an image that emphasizes the ball itself. The first event warning image 15 is an image that alerts the presence of an object that obstructs traveling ahead of the host vehicle, and an image that alerts the presence of a ball in the example of FIG. 6. It is. The primary event warning image 14 and the primary event warning image 15 are not limited to being displayed together as shown in FIG. 6, and either one may be displayed.
 ステップS14において、第2次事象推定部4は、第1次事象認識部3が認識した第1次事象から第2次事象を推定する。具体的には、第2次事象推定部4は、自身が有している図3に示すようなデータに基づいて第2次事象を推定する。ここでは、第2次事象推定部4は、「子どもの飛び出し」を第2次事象として推定する。 In step S14, the secondary event estimation unit 4 estimates the secondary event from the primary event recognized by the primary event recognition unit 3. Specifically, the secondary event estimation unit 4 estimates the secondary event based on data as shown in FIG. Here, the secondary event estimation unit 4 estimates “child's jumping out” as the secondary event.
 ステップS15において、警報画像生成部9は、第2次事象推定部4が推定した第2次事象を示す第2次事象警報画像を生成する。報知制御部5は、警報画像生成部9が生成した第2次事象警報画像をHUD11に表示する制御を行う。 In step S15, the warning image generation unit 9 generates a secondary event warning image indicating the secondary event estimated by the secondary event estimation unit 4. The notification control unit 5 performs control to display the second event warning image generated by the warning image generation unit 9 on the HUD 11.
 図7は、第2次事象警報画像をHUD11に表示した場合の一例を示す図である。図7において、第2次事象警報画像16は、自車両の前方に子どもが飛び出して来るような画像である。なお、第2次事象警報画像16は、予め定められた位置に表示される。第2次事象警報画像16を表示する位置は、ユーザが任意に設定してもよい。 FIG. 7 is a diagram showing an example when the second event warning image is displayed on the HUD 11. In FIG. 7, the secondary event warning image 16 is an image in which a child jumps out in front of the host vehicle. The secondary event warning image 16 is displayed at a predetermined position. The position at which the second event warning image 16 is displayed may be arbitrarily set by the user.
 以上のことから、本実施の形態1によれば、報知制御装置1,6は、実際の状況である第1次事象だけでなく、第1次事象後に生じ得る状況である第2次事象も報知しているため、ドライバに対して適切な報知を行うことができる。従って、ドライバは、第2次事象を考慮した適切な対応をとることができる。 From the above, according to the first embodiment, the notification control devices 1 and 6 not only provide the primary event that is the actual situation, but also the secondary event that can occur after the primary event. Since the notification is made, it is possible to appropriately notify the driver. Therefore, the driver can take an appropriate response in consideration of the secondary event.
 <実施の形態2>
 <構成>
 まず、本発明の実施の形態2による報知制御装置の構成について説明する。
<Embodiment 2>
<Configuration>
First, the configuration of the notification control apparatus according to Embodiment 2 of the present invention will be described.
 図8は、本実施の形態2による報知制御装置17の構成の一例を示すブロック図である。 FIG. 8 is a block diagram showing an example of the configuration of the notification control device 17 according to the second embodiment.
 図8に示すように、報知制御装置17は、第2次事象位置推定部18を備えることを特徴としている。その他の構成および動作は、実施の形態1と同様であるため、ここでは詳細な説明を省略する。 As shown in FIG. 8, the notification control device 17 is characterized by including a second event position estimation unit 18. Since other configurations and operations are the same as those in the first embodiment, detailed description thereof is omitted here.
 第2次事象位置推定部18は、第1次事象認識部3が第1次事象を認識した位置に基づいて、第2次事象推定部4が推定した第2次事象の発生位置を推定する。警報画像生成部9は、第2次事象位置推定部18が推定した発生位置に対応する第2次事象警報画像を生成する。 The secondary event position estimation unit 18 estimates the occurrence position of the secondary event estimated by the secondary event estimation unit 4 based on the position where the primary event recognition unit 3 has recognized the primary event. . The alarm image generation unit 9 generates a secondary event alarm image corresponding to the occurrence position estimated by the secondary event position estimation unit 18.
 報知制御装置17における第2次事象位置推定部18の機能は、処理回路により実現される。すなわち、第2次事象位置推定部18は、第2次事象の発生位置を推定するための処理回路を備える。処理回路は、例えば図4に示すような、メモリ13に格納されるプログラムを実行するプロセッサ12である。 The function of the secondary event position estimation unit 18 in the notification control device 17 is realized by a processing circuit. That is, the secondary event position estimation unit 18 includes a processing circuit for estimating the occurrence position of the secondary event. The processing circuit is, for example, a processor 12 that executes a program stored in the memory 13 as shown in FIG.
 報知制御装置17における第2次事象位置推定部18の機能は、ソフトウェア、ファームウェア、またはソフトウェアとファームウェアとの組み合わせにより実現される。ソフトウェアまたはファームウェアは、プログラムとして記述され、メモリ13に格納される。処理回路は、メモリ13に記憶されたプログラムを読み出して実行することにより、第2次事象位置推定部18の機能を実現する。すなわち、報知制御装置17は、第2次事象の発生位置を推定するステップが結果的に実行されることになるプログラムを格納するためのメモリ13を備える。また、このプログラムは、第2次事象位置推定部18の手順または方法をコンピュータに実行させるものであるともいえる。 The function of the secondary event position estimation unit 18 in the notification control device 17 is realized by software, firmware, or a combination of software and firmware. Software or firmware is described as a program and stored in the memory 13. The processing circuit implements the function of the secondary event position estimating unit 18 by reading and executing the program stored in the memory 13. That is, the notification control device 17 includes a memory 13 for storing a program in which the step of estimating the occurrence position of the second event is executed as a result. It can also be said that this program causes the computer to execute the procedure or method of the secondary event position estimation unit 18.
 <動作>
 次に、報知制御装置17の動作について説明する。
<Operation>
Next, the operation of the notification control device 17 will be described.
 図9は、報知制御装置17の動作の一例を示すフローチャートである。なお、図9のステップS21~ステップS24は、図5のステップS11~ステップS14に対応しているため、ここでは説明を省略する。以下では、ステップS25およびステップS26について説明する。周辺認識装置10はカメラであり、自車両の前方を撮影するものとする。 FIG. 9 is a flowchart showing an example of the operation of the notification control device 17. Note that steps S21 to S24 in FIG. 9 correspond to steps S11 to S14 in FIG. Below, step S25 and step S26 are demonstrated. The periphery recognition device 10 is a camera and shoots ahead of the host vehicle.
 ステップS25において、第2次事象位置推定部18は、周辺情報取得部2が取得した画像における、第1次事象認識部3が第1次事象を認識した位置付近の予め定められた位置を、第2次事象の発生位置と推定する。 In step S25, the secondary event position estimation unit 18 determines a predetermined position near the position where the primary event recognition unit 3 has recognized the primary event in the image acquired by the peripheral information acquisition unit 2. Presumed as the occurrence position of the secondary event.
 ステップS26において、警報画像生成部9は、第2次事象位置推定部18が推定した第2次事象の発生位置に対応する第2次事象警報画像を生成する。報知制御部5は、警報画像生成部9が生成した第2次事象警報画像をHUD11に表示する制御を行う。 In step S26, the alarm image generation unit 9 generates a secondary event alarm image corresponding to the occurrence position of the secondary event estimated by the secondary event position estimation unit 18. The notification control unit 5 performs control to display the second event warning image generated by the warning image generation unit 9 on the HUD 11.
 図10は、第2次事象警報画像をHUD11に表示した場合の一例を示す図である。図10において、第2次事象警報画像16は、第2次事象位置推定部18が推定した発生位置に表示される。図10の例では、子どもがボールを追って自車両の前方に飛び出して来る様子が表現されている。 FIG. 10 is a diagram illustrating an example when the second event warning image is displayed on the HUD 11. In FIG. 10, the secondary event warning image 16 is displayed at the occurrence position estimated by the secondary event position estimation unit 18. In the example of FIG. 10, a state in which a child follows the ball and jumps out ahead of the host vehicle is expressed.
 以上のことから、本実施の形態2によれば、報知制御装置17は、第2次事象が発生する可能性が高い位置に第2次事象を表示するため、ドライバは第2次事象を実施の形態1よりも正確に把握することができる。 From the above, according to the second embodiment, the notification control device 17 displays the second event at a position where the second event is highly likely to occur, so the driver performs the second event. It is possible to grasp more accurately than the first form.
 <実施の形態3>
 <構成>
 まず、本発明の実施の形態3による報知制御装置の構成について説明する。
<Embodiment 3>
<Configuration>
First, the configuration of the notification control apparatus according to Embodiment 3 of the present invention will be described.
 図11は、本実施の形態3による報知制御装置19の構成の一例を示すブロック図である。 FIG. 11 is a block diagram illustrating an example of the configuration of the notification control device 19 according to the third embodiment.
 図11に示すように、報知制御装置19は、認識範囲制限部20を備えることを特徴としている。その他の構成および動作は、実施の形態2と同様であるため、ここでは詳細な説明を省略する。 As shown in FIG. 11, the notification control device 19 includes a recognition range restriction unit 20. Since other configurations and operations are the same as those in the second embodiment, detailed description thereof is omitted here.
 認識範囲制限部20は、第1次事象認識部3による認識範囲を、第2次事象位置推定部18が推定した第2次事象の発生位置を含む予め定められた範囲に制限する。第1次事象認識部3は、認識範囲制限部20によって制限された範囲内において第1次事象の認識を行う。例えば、第1次事象認識部3が画像から第1次事象を認識する場合、第1次事象認識部3は、認識範囲制限部20によって制限された画像の範囲内において第1次事象の認識を行う。 The recognition range restriction unit 20 restricts the recognition range by the first event recognition unit 3 to a predetermined range including the occurrence position of the second event estimated by the second event position estimation unit 18. The primary event recognition unit 3 recognizes the primary event within the range restricted by the recognition range restriction unit 20. For example, when the primary event recognition unit 3 recognizes the primary event from the image, the primary event recognition unit 3 recognizes the primary event within the range of the image restricted by the recognition range restriction unit 20. I do.
 報知制御装置19における認識範囲制限部20の機能は、処理回路により実現される。すなわち、認識範囲制限部20は、第1次事象認識部3による認識範囲を、第2次事象位置推定部18が推定した第2次事象の発生位置を含む予め定められた範囲に制限するための処理回路を備える。処理回路は、例えば図4に示すような、メモリ13に格納されるプログラムを実行するプロセッサ12である。 The function of the recognition range restriction unit 20 in the notification control device 19 is realized by a processing circuit. That is, the recognition range restriction unit 20 restricts the recognition range by the primary event recognition unit 3 to a predetermined range including the occurrence position of the secondary event estimated by the secondary event position estimation unit 18. The processing circuit is provided. The processing circuit is, for example, a processor 12 that executes a program stored in the memory 13 as shown in FIG.
 報知制御装置19における認識範囲制限部20の機能は、ソフトウェア、ファームウェア、またはソフトウェアとファームウェアとの組み合わせにより実現される。ソフトウェアまたはファームウェアは、プログラムとして記述され、メモリ13に格納される。処理回路は、メモリ13に記憶されたプログラムを読み出して実行することにより、認識範囲制限部20の機能を実現する。すなわち、報知制御装置19は、第1次事象認識部3による認識範囲を、第2次事象位置推定部18が推定した第2次事象の発生位置を含む予め定められた範囲に制限するステップが結果的に実行されることになるプログラムを格納するためのメモリ13を備える。また、このプログラムは、認識範囲制限部20の手順または方法をコンピュータに実行させるものであるともいえる。 The function of the recognition range restriction unit 20 in the notification control device 19 is realized by software, firmware, or a combination of software and firmware. Software or firmware is described as a program and stored in the memory 13. The processing circuit implements the function of the recognition range restriction unit 20 by reading and executing the program stored in the memory 13. That is, the notification control device 19 includes a step of limiting the recognition range by the primary event recognition unit 3 to a predetermined range including the occurrence position of the secondary event estimated by the secondary event position estimation unit 18. A memory 13 is provided for storing a program to be executed as a result. It can also be said that this program causes a computer to execute the procedure or method of the recognition range restriction unit 20.
 <動作>
 次に、報知制御装置19の動作について説明する。
<Operation>
Next, the operation of the notification control device 19 will be described.
 図12は、報知制御装置19の動作の一例を示すフローチャートである。なお、図12のステップS31~ステップS36は、図9のステップS21~ステップS26に対応しているため、ここでは説明を省略する。以下では、ステップS37およびステップS38について説明する。周辺認識装置10はカメラであり、自車両の前方を撮影するものとする。 FIG. 12 is a flowchart showing an example of the operation of the notification control device 19. Note that steps S31 to S36 in FIG. 12 correspond to steps S21 to S26 in FIG. Below, step S37 and step S38 are demonstrated. The periphery recognition device 10 is a camera and shoots ahead of the host vehicle.
 ステップS37において、報知制御装置19は、第1次事象認識部3による認識範囲を制限するか否かの判断を行う。第1次事象認識部3による認識範囲を制限するか否かの判断は、周辺情報取得部2が取得した画像の解像度に応じて行ってもよい。具体的には、画像の解像度が高い場合は、第1次事象認識部3における処理負荷が大きくなるため、第1次事象認識部3による認識範囲を制限すると判断してもよい。画像の解像度が低い場合は、第1次事象認識部3における処理負荷が小さいため、第1次事象認識部3による認識範囲を制限しないと判断してもよい。なお、第1次事象認識部3による認識範囲を制限するか否かの判断は、ユーザが任意に設定してもよい。第1次事象認識部3による認識範囲を制限する場合は、ステップS38に移行する。一方、第1次事象認識部3による認識範囲を制限しない場合は、ステップS31に移行する。 In step S37, the notification control device 19 determines whether or not to limit the recognition range by the primary event recognition unit 3. The determination of whether or not to limit the recognition range by the primary event recognition unit 3 may be performed according to the resolution of the image acquired by the peripheral information acquisition unit 2. Specifically, when the resolution of the image is high, the processing load on the primary event recognition unit 3 increases, and therefore it may be determined that the recognition range by the primary event recognition unit 3 is limited. When the resolution of the image is low, it may be determined that the recognition range by the primary event recognition unit 3 is not limited because the processing load on the primary event recognition unit 3 is small. The determination as to whether or not to limit the recognition range by the primary event recognition unit 3 may be arbitrarily set by the user. When limiting the recognition range by the primary event recognition part 3, it transfers to step S38. On the other hand, when the recognition range by the primary event recognition unit 3 is not limited, the process proceeds to step S31.
 ステップS38において、認識範囲制限部20は、第1次事象認識部3による認識範囲を、第2次事象位置推定部18が推定した第2次事象の発生位置を含む予め定められた範囲に制限する。具体的には、認識範囲制限部20は、第1次事象認識部3による認識範囲を、例えば図13に示すような認識範囲21に制限する。なお、図13に示す認識範囲21は、説明のために図示したものであり、実際にはHUD11に表示されない。 In step S38, the recognition range restriction unit 20 restricts the recognition range by the primary event recognition unit 3 to a predetermined range including the occurrence position of the secondary event estimated by the secondary event position estimation unit 18. To do. Specifically, the recognition range restriction unit 20 restricts the recognition range by the primary event recognition unit 3 to, for example, a recognition range 21 as shown in FIG. Note that the recognition range 21 shown in FIG. 13 is shown for explanation, and is not actually displayed on the HUD 11.
 以上のことから、本実施の形態3によれば、第1次事象認識部3による認識範囲を制限するため、第1次事象認識部3による第1次事象を認識する効率を向上させることができ、第2次事象を早く表示することができる。また、第1次事象認識部3によって第1次事象を認識するためのリソースが限られた場合であっても、第1次事象認識部3は第1次事象の認識をスムーズに行うことができる。 From the above, according to the third embodiment, since the recognition range by the primary event recognition unit 3 is limited, the efficiency of recognizing the primary event by the primary event recognition unit 3 can be improved. The secondary event can be displayed quickly. Moreover, even if the resources for recognizing the primary event by the primary event recognition unit 3 are limited, the primary event recognition unit 3 can smoothly recognize the primary event. it can.
 なお、図12の例では、第1次事象認識部3が、周辺情報取得部2が取得する全ての画像に対して認識範囲制限部20が制限した範囲内で第1次事象を認識する場合について説明したが、これに限るものではない。例えば、第1次事象認識部3は、認識範囲制限部20が制限した範囲内については、全ての画像に対して第1次事象の認識を行い、認識範囲制限部20が制限した範囲外については、予め定められた間隔ごとに周辺情報取得部2が取得した画像に対して第1次事象の認識を行うようにしてもよい。 In the example of FIG. 12, the primary event recognition unit 3 recognizes the primary event within the range limited by the recognition range limitation unit 20 for all images acquired by the peripheral information acquisition unit 2. However, the present invention is not limited to this. For example, the primary event recognition unit 3 recognizes the primary event for all images within the range restricted by the recognition range restriction unit 20 and is outside the range restricted by the recognition range restriction unit 20. May recognize the primary event for the image acquired by the peripheral information acquisition unit 2 at predetermined intervals.
 <実施の形態4>
 <構成>
 まず、本発明の実施の形態4による報知制御装置の構成について説明する。
<Embodiment 4>
<Configuration>
First, the configuration of a notification control apparatus according to Embodiment 4 of the present invention will be described.
 図14は、本実施の形態4による報知制御装置22の構成の一例を示すブロック図である。 FIG. 14 is a block diagram showing an example of the configuration of the notification control device 22 according to the fourth embodiment.
 図14に示すように、報知制御装置22は、通信部23を備えることを特徴としている。その他の構成および動作は、実施の形態3と同様であるため、ここでは詳細な説明を省略する。 As shown in FIG. 14, the notification control device 22 is characterized by including a communication unit 23. Other configurations and operations are the same as those in the third embodiment, and thus detailed description thereof is omitted here.
 通信部23は、報知制御装置22の外部に設けられた画像認識学習サーバ24および第2次事象推定データサーバ25の各々と通信可能に接続されており、画像認識学習サーバ24および第2次事象推定データサーバ25の各々から後述するパラメータおよびデータを取得する。 The communication unit 23 is communicably connected to each of the image recognition learning server 24 and the second event estimation data server 25 provided outside the notification control device 22, and the image recognition learning server 24 and the second event. Parameters and data to be described later are acquired from each of the estimated data servers 25.
 画像認識学習サーバ24は、第1次事象認識部3が画像から第1次事象を認識する際に使用するパラメータについて、第1次事象認識部3による第1次事象の認識精度が向上するような機械学習を行う。第2次事象推定データサーバ25は、第2次事象推定部4が第2次事象を推定する際に使用するデータを蓄積する。なお、画像認識学習サーバ24および第2次事象推定データサーバ25は、別個に設けられてもよく、一体に設けられてもよい。 The image recognition learning server 24 improves the recognition accuracy of the primary event by the primary event recognition unit 3 for parameters used when the primary event recognition unit 3 recognizes the primary event from the image. Machine learning. The secondary event estimation data server 25 accumulates data used when the secondary event estimation unit 4 estimates the secondary event. The image recognition learning server 24 and the second event estimation data server 25 may be provided separately or may be provided integrally.
 報知制御装置22における通信部23の機能は、処理回路により実現される。すなわち、通信部23は、画像認識学習サーバ24および第2次事象推定データサーバ25の各々と通信するための処理回路を備える。処理回路は、例えば図4に示すような、メモリ13に格納されるプログラムを実行するプロセッサ12である。 The function of the communication unit 23 in the notification control device 22 is realized by a processing circuit. That is, the communication unit 23 includes a processing circuit for communicating with each of the image recognition learning server 24 and the second event estimation data server 25. The processing circuit is, for example, a processor 12 that executes a program stored in the memory 13 as shown in FIG.
 報知制御装置22における通信部23の機能は、ソフトウェア、ファームウェア、またはソフトウェアとファームウェアとの組み合わせにより実現される。ソフトウェアまたはファームウェアは、プログラムとして記述され、メモリ13に格納される。処理回路は、メモリ13に記憶されたプログラムを読み出して実行することにより、通信部23の機能を実現する。すなわち、報知制御装置22は、画像認識学習サーバ24および第2次事象推定データサーバ25の各々と通信するステップが結果的に実行されることになるプログラムを格納するためのメモリ13を備える。また、このプログラムは、通信部23の手順または方法をコンピュータに実行させるものであるともいえる。 The function of the communication unit 23 in the notification control device 22 is realized by software, firmware, or a combination of software and firmware. Software or firmware is described as a program and stored in the memory 13. The processing circuit implements the function of the communication unit 23 by reading and executing the program stored in the memory 13. That is, the notification control device 22 includes a memory 13 for storing a program in which a step of communicating with each of the image recognition learning server 24 and the second event estimation data server 25 is executed as a result. It can also be said that this program causes a computer to execute the procedure or method of the communication unit 23.
 <動作>
 次に、報知制御装置22の動作について説明する。
<Operation>
Next, the operation of the notification control device 22 will be described.
 図15は、報知制御装置22の動作の一例を示すフローチャートである。なお、図15のステップS41~ステップS48は、図12のステップS31~ステップS38に対応しているため、ここでは説明を省略する。なお、図15のステップS47において、第1次事象認識部3による認識範囲を制限しない場合は、ステップS49に移行する。以下では、ステップS49について説明する。周辺認識装置10はカメラであり、自車両の前方を撮影するものとする。 FIG. 15 is a flowchart showing an example of the operation of the notification control device 22. Note that steps S41 to S48 in FIG. 15 correspond to steps S31 to S38 in FIG. In step S47 of FIG. 15, if the recognition range by the primary event recognition unit 3 is not limited, the process proceeds to step S49. Hereinafter, step S49 will be described. The periphery recognition device 10 is a camera and shoots ahead of the host vehicle.
 ステップS49において、通信部23は、第1次事象認識部3が画像から第1次事象を認識する際に使用するパラメータを画像認識学習サーバ24から取得する。第1次事象認識部3が使用するパラメータは、通信部23が取得したパラメータに更新される。また、通信部23は、第2次事象推定部4が第2次事象を推定する際に使用するデータを第2次事象推定データサーバ25から取得する。第2次事象推定部4が使用するデータは、通信部23が取得したデータに更新される。なお、ステップS49の処理は、ステップS47またはステップS48の後に行うことに限らず、任意のタイミングであってもよい。 In step S49, the communication unit 23 acquires, from the image recognition learning server 24, parameters used when the primary event recognition unit 3 recognizes the primary event from the image. The parameter used by the primary event recognition unit 3 is updated to the parameter acquired by the communication unit 23. Further, the communication unit 23 acquires data used when the secondary event estimation unit 4 estimates the secondary event from the secondary event estimation data server 25. The data used by the second event estimation unit 4 is updated to the data acquired by the communication unit 23. Note that the process of step S49 is not limited to being performed after step S47 or step S48, and may be performed at an arbitrary timing.
 次に、画像認識学習サーバ24および第2次事象推定データサーバ25の動作について説明する。 Next, operations of the image recognition learning server 24 and the secondary event estimation data server 25 will be described.
 図16は、画像認識学習サーバ24および第2次事象推定データサーバ25の動作の一例を示すフローチャートである。 FIG. 16 is a flowchart showing an example of operations of the image recognition learning server 24 and the secondary event estimation data server 25.
 ステップS51において、画像認識学習サーバ24は、第1次事象認識部3が画像から第1次事象を認識する際に使用するパラメータについて、第1次事象認識部3による第1次事象の認識精度が向上するような機械学習を行う。機械学習されたパラメータは更新される。 In step S51, the image recognition learning server 24 recognizes the primary event recognition accuracy by the primary event recognition unit 3 for parameters used when the primary event recognition unit 3 recognizes the primary event from the image. Do machine learning that will improve. Machine-learned parameters are updated.
 ステップS52において、第2次事象推定データサーバ25は、第2次事象推定部4が第2次事象を推定する際に使用するデータを蓄積する。具体的には、第2次事象推定データサーバ25は、例えば図3に示すような第1次事象と第2次事象とを対応付けたデータを蓄積し、必要に応じて更新する。 In step S52, the secondary event estimation data server 25 accumulates data used when the secondary event estimation unit 4 estimates the secondary event. Specifically, the secondary event estimation data server 25 accumulates data associating the primary event and the secondary event as shown in FIG. 3, for example, and updates it as necessary.
 ステップS53において、画像認識学習サーバ24は、第1次事象認識部3が画像から第1次事象を認識する際に使用するパラメータを、報知制御装置22の通信部23に送信する。また、第2次事象推定データサーバ25は、第2次事象推定部4が第2次事象を推定する際に使用するデータを、報知制御装置22の通信部23に送信する。なお、画像認識学習サーバ24および第2次事象推定データサーバ25は、パラメータおよびデータを同時に送信してもよく、別々に送信してもよい。また、画像認識学習サーバ24および第2次事象推定データサーバ25がパラメータおよびデータを送信するタイミングは、予め定められたタイミングであってもよく、ユーザが設定した任意のタイミングであってもよい。 In step S <b> 53, the image recognition learning server 24 transmits parameters used when the primary event recognition unit 3 recognizes the primary event from the image to the communication unit 23 of the notification control device 22. Further, the secondary event estimation data server 25 transmits data used when the secondary event estimation unit 4 estimates the secondary event to the communication unit 23 of the notification control device 22. Note that the image recognition learning server 24 and the second event estimation data server 25 may transmit parameters and data simultaneously or separately. Further, the timing at which the image recognition learning server 24 and the second event estimation data server 25 transmit parameters and data may be a predetermined timing or an arbitrary timing set by the user.
 以上のことから、本実施の形態4によれば、第1次事象認識部3が画像から第1次事象を認識する際に使用するパラメータを更新することによって、第1次事象認識部3による第1次事象の認識精度を向上させることが可能となる。また、第2次事象推定部4が第2次事象を推定する際に使用するデータを更新することによって、第2次事象推定部4が第2次事象を推定する精度を向上させることができる。 From the above, according to the fourth embodiment, the primary event recognition unit 3 updates the parameters used when the primary event recognition unit 3 recognizes the primary event from the image, so that the primary event recognition unit 3 It becomes possible to improve the recognition accuracy of the primary event. Moreover, the accuracy which the secondary event estimation part 4 estimates a secondary event can be improved by updating the data used when the secondary event estimation part 4 estimates a secondary event. .
 なお、図14では、図11に示す報知制御装置19に通信部23をさらに備える場合について説明したが、これに限るものではない。例えば、図2に示す報知制御装置6、および図8に示す報知制御装置17の各々に通信部23を備えるようにしてもよい。 In addition, although FIG. 14 demonstrated the case where the notification control apparatus 19 shown in FIG. 11 was further equipped with the communication part 23, it is not restricted to this. For example, the notification control device 6 shown in FIG. 2 and the notification control device 17 shown in FIG.
 <実施の形態5>
 <構成>
 まず、本発明の実施の形態5による報知制御装置の構成について説明する。
<Embodiment 5>
<Configuration>
First, the configuration of the notification control apparatus according to Embodiment 5 of the present invention will be described.
 図17は、本実施の形態5による報知制御装置26の構成の一例を示すブロック図である。 FIG. 17 is a block diagram showing an example of the configuration of the notification control device 26 according to the fifth embodiment.
 図17に示すように、報知制御装置26は、車両制御部27を備えることを特徴としている。その他の構成および動作は、実施の形態1と同様であるため、ここでは詳細な説明を省略する。 As shown in FIG. 17, the notification control device 26 includes a vehicle control unit 27. Since other configurations and operations are the same as those in the first embodiment, detailed description thereof is omitted here.
 車両制御部27は、第2次事象推定部4による第2次事象の推定結果に基づいて自車両の運転操作を制御する。具体的には、車両制御部27は、自車両のアクセル、ブレーキ、およびハンドルの各操作を制御する。 The vehicle control unit 27 controls the driving operation of the host vehicle based on the secondary event estimation result by the secondary event estimation unit 4. Specifically, the vehicle control unit 27 controls each operation of the accelerator, the brake, and the handle of the host vehicle.
 報知制御装置26における車両制御部27の機能は、処理回路により実現される。すなわち、車両制御部27は、第2次事象推定部4による第2次事象の推定結果に基づいて自車両の運転を制御するための処理回路を備える。処理回路は、例えば図4に示すような、メモリ13に格納されるプログラムを実行するプロセッサ12である。 The function of the vehicle control unit 27 in the notification control device 26 is realized by a processing circuit. That is, the vehicle control unit 27 includes a processing circuit for controlling the driving of the host vehicle based on the secondary event estimation result by the secondary event estimation unit 4. The processing circuit is, for example, a processor 12 that executes a program stored in the memory 13 as shown in FIG.
 報知制御装置26における車両制御部27の機能は、ソフトウェア、ファームウェア、またはソフトウェアとファームウェアとの組み合わせにより実現される。ソフトウェアまたはファームウェアは、プログラムとして記述され、メモリ13に格納される。処理回路は、メモリ13に記憶されたプログラムを読み出して実行することにより、車両制御部27の機能を実現する。すなわち、車両制御部27は、第2次事象推定部4による第2次事象の推定結果に基づいて自車両の運転を制御するステップが結果的に実行されることになるプログラムを格納するためのメモリ13を備える。また、このプログラムは、車両制御部27の手順または方法をコンピュータに実行させるものであるともいえる。 The function of the vehicle control unit 27 in the notification control device 26 is realized by software, firmware, or a combination of software and firmware. Software or firmware is described as a program and stored in the memory 13. The processing circuit implements the function of the vehicle control unit 27 by reading and executing the program stored in the memory 13. That is, the vehicle control unit 27 stores a program in which the step of controlling the driving of the host vehicle is executed as a result based on the estimation result of the secondary event by the secondary event estimation unit 4. A memory 13 is provided. It can also be said that this program causes a computer to execute the procedure or method of the vehicle control unit 27.
 <動作>
 次に、報知制御装置26の動作について説明する。
<Operation>
Next, the operation of the notification control device 26 will be described.
 図18は、報知制御装置26の動作の一例を示すフローチャートである。なお、図18のステップS61~ステップS65は、図5のステップS11~ステップS15に対応しているため、ここでは説明を省略する。以下では、ステップS66について説明する。周辺認識装置10はカメラであり、自車両の前方を撮影するものとする。 FIG. 18 is a flowchart showing an example of the operation of the notification control device 26. Note that steps S61 to S65 in FIG. 18 correspond to steps S11 to S15 in FIG. Hereinafter, step S66 will be described. The periphery recognition device 10 is a camera and shoots ahead of the host vehicle.
 ステップS66において、車両制御部27は、第2次事象推定部4による第2次事象の推定結果に基づいて自車両の運転操作を制御する。具体的には、車両制御部27は、第2次事象を回避する方向のハンドルの反力を小さくする、第2次事象を回避する方向にハンドル操作する、ブレーキを早目にかける等といった自車両の運転操作を制御する。例えば図3を参照して、第2次事象が「割り込み車両」である場合、車両制御部27は、自車両を減速する制御を行う。 In step S66, the vehicle control unit 27 controls the driving operation of the host vehicle based on the secondary event estimation result by the secondary event estimation unit 4. Specifically, the vehicle control unit 27 reduces the reaction force of the steering wheel in the direction of avoiding the secondary event, operates the steering wheel in the direction of avoiding the secondary event, applies the brake early, etc. Control the driving operation of the vehicle. For example, referring to FIG. 3, when the second event is “interrupt vehicle”, vehicle control unit 27 performs control to decelerate the host vehicle.
 なお、車両制御部27は、第1次事象認識部3による認識結果に基づいて自車両の運転操作を制御してもよい。例えば図3を参照して、第1次事象が「ボールの飛び出し」である場合、車両制御部27は、ボールを回避するようなハンドル操作となるように制御してもよい。 The vehicle control unit 27 may control the driving operation of the host vehicle based on the recognition result by the primary event recognition unit 3. For example, referring to FIG. 3, when the first event is “ball jumping out”, the vehicle control unit 27 may perform control so that the steering wheel operation avoids the ball.
 以上のことから、本実施の形態5によれば、第2次事象を表示するだけでなく、当該第2次事象を回避するように自車両の運転操作を制御している。従って、報知制御装置26は、第2次事象を報知するだけでなく、自車両の運転操作を支援することができる。 From the above, according to the fifth embodiment, not only the secondary event is displayed, but also the driving operation of the host vehicle is controlled so as to avoid the secondary event. Therefore, the notification control device 26 can not only notify the second event but also support the driving operation of the host vehicle.
 なお、図17では、図2に示す報知制御装置6に車両制御部27をさらに備える場合について説明したが、これに限るものではない。例えば、図8に示す報知制御装置17、図11に示す報知制御装置19、および図14に示す報知制御装置22の各々に車両制御部27を備えるようにしてもよい。 In addition, although FIG. 17 demonstrated the case where the notification control apparatus 6 shown in FIG. 2 was further provided with the vehicle control part 27, it is not restricted to this. For example, the vehicle control unit 27 may be provided in each of the notification control device 17 shown in FIG. 8, the notification control device 19 shown in FIG. 11, and the notification control device 22 shown in FIG.
 実施の形態1~5では、第1次事象認識部3が第1次事象を認識する場合について説明したが、これに限るものではない。例えば、他車両が認識した第1次事象を、当該他車両から車々間通信で取得してもよい。また、道路に設置されたカメラなどのインフラ設備が認識した第1次事象を、当該インフラ設備から路車間通信で取得してもよい。この場合、自車両では、当該自車両に設けられた周辺認識装置10では撮影できない場所に関する第1次事象を用いて第2次事象を推定することができる。 In Embodiments 1 to 5, the case where the primary event recognition unit 3 recognizes the primary event has been described, but the present invention is not limited to this. For example, the first event recognized by another vehicle may be acquired from the other vehicle by inter-vehicle communication. Moreover, you may acquire the 1st event recognized by infrastructure facilities, such as a camera installed in the road, from the said infrastructure facility by road-to-vehicle communication. In this case, the host vehicle can estimate the second event using the first event regarding the place that cannot be captured by the periphery recognition device 10 provided in the host vehicle.
 実施の形態1~5において、第2次事象推定部4が推定した第2次事象は、他車両に送信してもよい。 In Embodiments 1 to 5, the secondary event estimated by the secondary event estimation unit 4 may be transmitted to another vehicle.
 実施の形態1~5では、周辺認識装置10であるカメラが自車両の前方を撮影する場合について説明したが、これに限るものではない。周辺認識装置10であるカメラは、自車両の周囲の全方向または任意の方向を撮影するようにしてもよい。 In Embodiments 1 to 5, the case where the camera that is the periphery recognition device 10 captures the front of the host vehicle has been described, but the present invention is not limited to this. The camera that is the periphery recognition device 10 may capture all directions around the host vehicle or any direction.
 以上で説明した報知制御装置は、車載用ナビゲーション装置、すなわちカーナビゲーション装置だけでなく、車両に搭載可能なPND(Portable Navigation Device)およびサーバなどを適宜に組み合わせてシステムとして構築されるナビゲーション装置あるいはナビゲーション装置以外の装置にも適用することができる。この場合、報知制御装置の各機能あるいは各構成要素は、上記システムを構築する各機能に分散して配置される。 The notification control device described above is not only a vehicle navigation device, that is, a car navigation device, but also a navigation device or navigation constructed as a system by appropriately combining a PND (Portable Navigation Device) and a server that can be mounted on a vehicle. The present invention can be applied to apparatuses other than the apparatus. In this case, each function or each component of the notification control device is distributed and arranged in each function that constructs the system.
 具体的には、一例として、報知制御装置の機能をサーバに配置することができる。例えば、図19に示すように、車両側に周辺認識装置10およびHUD11を備え、サーバ28に周辺情報取得部2、第1次事象認識部3、第2次事象推定部4、報知制御部5、自車両情報取得部7、地図情報取得部8、および警報画像生成部9を備えることによって、報知制御システムを構築することができる。図8に示す報知制御装置17、図11に示す報知制御装置19、図14に示す報知制御装置22、および図17に示す報知制御装置26についても同様である。 Specifically, as an example, the function of the notification control device can be arranged in the server. For example, as shown in FIG. 19, the vehicle side includes a peripheral recognition device 10 and a HUD 11, and the server 28 includes a peripheral information acquisition unit 2, a primary event recognition unit 3, a secondary event estimation unit 4, and a notification control unit 5. By providing the own vehicle information acquisition unit 7, the map information acquisition unit 8, and the alarm image generation unit 9, a notification control system can be constructed. The same applies to the notification control device 17 shown in FIG. 8, the notification control device 19 shown in FIG. 11, the notification control device 22 shown in FIG. 14, and the notification control device 26 shown in FIG.
 また、報知制御装置の機能をサーバおよび携帯通信端末30に配置することができる。例えば、図20に示すように、車両側に周辺認識装置10およびHUD11を備え、サーバ29に周辺情報取得部2、第1次事象認識部3、第2次事象推定部4、自車両情報取得部7、地図情報取得部8、および警報画像生成部9を備え、携帯通信端末30に報知制御部5を備えることによって、報知制御システムを構築することができる。図8に示す報知制御装置17、図11に示す報知制御装置19、図14に示す報知制御装置22、および図17に示す報知制御装置26についても同様である。 Further, the function of the notification control device can be arranged in the server and the mobile communication terminal 30. For example, as shown in FIG. 20, the vehicle is provided with the periphery recognition device 10 and the HUD 11, and the server 29 includes the periphery information acquisition unit 2, the primary event recognition unit 3, the secondary event estimation unit 4, and the own vehicle information acquisition. By providing the unit 7, the map information acquisition unit 8, and the alarm image generation unit 9, and the notification control unit 5 in the mobile communication terminal 30, a notification control system can be constructed. The same applies to the notification control device 17 shown in FIG. 8, the notification control device 19 shown in FIG. 11, the notification control device 22 shown in FIG. 14, and the notification control device 26 shown in FIG.
 上記の構成とした場合であっても、上記の実施の形態と同様の効果が得られる。 Even in the case of the above configuration, the same effect as the above embodiment can be obtained.
 また、上記の実施の形態における動作を実行するソフトウェア(報知制御方法)を、例えばサーバまたは携帯通信端末に組み込んでもよい。 Further, software (notification control method) for executing the operation in the above embodiment may be incorporated in, for example, a server or a mobile communication terminal.
 具体的には、一例として、上記の報知制御方法は、少なくとも自車両の周辺の状況を示す周辺情報を取得し、取得した周辺情報から、自車両の周辺に存在する物体または人物の状況である第1次事象を認識し、認識した第1次事象の後に生じ得る状況であって、第1次事象の状況と因果関係を有する状況である第2次事象を、予め定められた規則により推定し、推定した第2次事象を報知する制御を行う。 Specifically, as an example, the above-described notification control method acquires at least peripheral information indicating a situation around the host vehicle, and is a situation of an object or a person existing around the host vehicle from the acquired peripheral information. Recognizing a primary event and estimating a secondary event that can occur after the recognized primary event and has a causal relationship with the situation of the primary event according to a predetermined rule Then, control for notifying the estimated secondary event is performed.
 上記より、上記の実施の形態における動作を実行するソフトウェアをサーバまたは携帯通信端末に組み込んで動作させることによって、上記の実施の形態と同様の効果が得られる。 From the above, the same effect as in the above embodiment can be obtained by operating the software that executes the operation in the above embodiment in a server or a portable communication terminal.
 なお、本発明は、その発明の範囲内において、各実施の形態を自由に組み合わせたり、各実施の形態を適宜、変形、省略することが可能である。 It should be noted that the present invention can be freely combined with each other within the scope of the invention, and each embodiment can be appropriately modified or omitted.
 本発明は詳細に説明されたが、上記した説明は、すべての態様において、例示であって、この発明がそれに限定されるものではない。例示されていない無数の変形例が、この発明の範囲から外れることなく想定され得るものと解される。 Although the present invention has been described in detail, the above description is illustrative in all aspects, and the present invention is not limited thereto. It is understood that countless variations that are not illustrated can be envisaged without departing from the scope of the present invention.
 1 報知制御装置、2 周辺情報取得部、3 第1次事象認識部、4 第2次事象推定部、5 報知制御部、6 報知制御装置、7 自車両情報取得部、8 地図情報取得部、9 警報画像生成部、10 周辺認識装置、11 HUD、12 プロセッサ、13 メモリ、14,15 第1次事象警報画像、16 第2次事象警報画像、17 報知制御装置、18 第2次事象位置推定部、19 報知制御装置、20 認識範囲制限部、21 認識範囲、22 報知制御装置、23 通信部、24 画像認識学習サーバ、25 第2次事象推定データサーバ、26 報知制御装置、27 車両制御部、28,29 サーバ、30 携帯通信端末。 1 notification control device, 2 surrounding information acquisition unit, 3 primary event recognition unit, 4 secondary event estimation unit, 5 notification control unit, 6 notification control device, 7 own vehicle information acquisition unit, 8 map information acquisition unit, 9 alarm image generation unit, 10 peripheral recognition device, 11 HUD, 12 processor, 13 memory, 14, 15 primary event warning image, 16 secondary event warning image, 17 notification control device, 18 secondary event position estimation Unit, 19 notification control device, 20 recognition range restriction unit, 21 recognition range, 22 notification control device, 23 communication unit, 24 image recognition learning server, 25 secondary event estimation data server, 26 notification control device, 27 vehicle control unit 28, 29 servers, 30 mobile communication terminals.

Claims (8)

  1.  少なくとも自車両の周辺の状況を示す周辺情報を取得する周辺情報取得部と、
     前記周辺情報取得部が取得した前記周辺情報から、前記自車両の周辺に存在する物体または人物の状況である第1次事象を認識する第1次事象認識部と、
     前記第1次事象認識部が認識した前記第1次事象の後に生じ得る状況であって、前記第1次事象の前記状況と因果関係を有する状況である第2次事象を、予め定められた規則により推定する第2次事象推定部と、
     前記第2次事象推定部が推定した前記第2次事象を報知する制御を行う報知制御部と、
    を備える、報知制御装置。
    A peripheral information acquisition unit that acquires peripheral information indicating at least the situation around the host vehicle;
    A primary event recognition unit for recognizing a primary event that is a situation of an object or person existing around the host vehicle from the peripheral information acquired by the peripheral information acquisition unit;
    A secondary event that is a situation that can occur after the primary event recognized by the primary event recognition unit and has a causal relationship with the situation of the primary event is predetermined. A secondary event estimator that estimates by rules;
    A notification control unit that performs control to notify the second event estimated by the second event estimation unit;
    A notification control device comprising:
  2.  前記第2次事象推定部が推定した前記第2次事象を示す第2次事象警報画像を生成する警報画像生成部をさらに備え、
     前記報知制御部は、前記警報画像生成部が生成した前記第2次事象警報画像を報知する制御を行うことを特徴とする、請求項1に記載の報知制御装置。
    An alarm image generation unit that generates a secondary event alarm image indicating the secondary event estimated by the secondary event estimation unit;
    The notification control apparatus according to claim 1, wherein the notification control unit performs control to notify the second event alarm image generated by the alarm image generation unit.
  3.  前記第1次事象認識部が前記第1次事象を認識した位置に基づいて、前記第2次事象の発生位置を推定する第2次事象位置推定部をさらに備え、
     前記警報画像生成部は、前記第2次事象位置推定部が推定した前記発生位置に対応する前記第2次事象警報画像を生成することを特徴とする、請求項2に記載の報知制御装置。
    A second event position estimating unit for estimating an occurrence position of the second event based on a position at which the first event recognizing unit has recognized the first event;
    The notification control device according to claim 2, wherein the warning image generation unit generates the second event warning image corresponding to the generation position estimated by the second event position estimation unit.
  4.  前記第1次事象認識部による認識範囲を、前記第2次事象位置推定部が推定した前記発生位置を含む予め定められた範囲に制限する認識範囲制限部をさらに備えることを特徴とする、請求項3に記載の報知制御装置。 The system further comprises a recognition range limiting unit that limits a recognition range by the first event recognition unit to a predetermined range including the occurrence position estimated by the second event position estimation unit. Item 6. The notification control device according to item 3.
  5.  前記第1次事象認識部が前記第1次事象を認識する際に使用するパラメータは、更新可能であることを特徴とする、請求項1に記載の報知制御装置。 The notification control device according to claim 1, wherein the parameter used when the first event recognition unit recognizes the first event can be updated.
  6.  前記第2次事象推定部が前記第2次事象を推定する際に使用するデータは、更新可能であることを特徴とする、請求項1に記載の報知制御装置。 The notification control apparatus according to claim 1, wherein the data used when the secondary event estimation unit estimates the secondary event can be updated.
  7.  前記第2次事象推定部による前記第2次事象の推定結果に基づいて、前記自車両の運転操作を制御する車両制御部をさらに備えることを特徴とする、請求項1に記載の報知制御装置。 The notification control device according to claim 1, further comprising a vehicle control unit that controls a driving operation of the host vehicle based on an estimation result of the secondary event by the secondary event estimation unit. .
  8.  少なくとも自車両の周辺の状況を示す周辺情報を取得し、
     前記取得した前記周辺情報から、前記自車両の周辺に存在する物体または人物の状況である第1次事象を認識し、
     前記認識した前記第1次事象の後に生じ得る状況であって、前記第1次事象の前記状況と因果関係を有する状況である第2次事象を、予め定められた規則により推定し、
     前記推定した前記第2次事象を報知する制御を行う、報知制御方法。
    Get at least peripheral information that shows the situation around your vehicle,
    Recognizing a first event which is a situation of an object or person existing around the vehicle from the acquired surrounding information,
    Estimating a secondary event that may occur after the recognized primary event and has a causal relationship with the situation of the primary event according to a predetermined rule;
    A notification control method for performing control to notify the estimated second event.
PCT/JP2017/000193 2017-01-06 2017-01-06 Reporting control device and reporting control method WO2018127962A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
JP2018560292A JP6661796B2 (en) 2017-01-06 2017-01-06 Notification control device and notification control method
PCT/JP2017/000193 WO2018127962A1 (en) 2017-01-06 2017-01-06 Reporting control device and reporting control method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2017/000193 WO2018127962A1 (en) 2017-01-06 2017-01-06 Reporting control device and reporting control method

Publications (1)

Publication Number Publication Date
WO2018127962A1 true WO2018127962A1 (en) 2018-07-12

Family

ID=62791235

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2017/000193 WO2018127962A1 (en) 2017-01-06 2017-01-06 Reporting control device and reporting control method

Country Status (2)

Country Link
JP (1) JP6661796B2 (en)
WO (1) WO2018127962A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112241677A (en) * 2019-07-17 2021-01-19 本田技研工业株式会社 Information providing device, information providing method, and storage medium
US20220126818A1 (en) * 2020-10-28 2022-04-28 Toyota Research Institute, Inc. Systems and methods for identifying high-risk driving situations from driving data

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2014021421A1 (en) * 2012-08-03 2014-02-06 クラリオン株式会社 Camera parameter computation device, navigation system, and camera parameter computation method
JP2014203349A (en) * 2013-04-08 2014-10-27 スズキ株式会社 Vehicle drive support device
JP2016192104A (en) * 2015-03-31 2016-11-10 キヤノンマーケティングジャパン株式会社 Information processor, processing method for the same, and program

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2006072830A (en) * 2004-09-03 2006-03-16 Aisin Aw Co Ltd Operation supporting system and operation supporting module
JP2006259895A (en) * 2005-03-15 2006-09-28 Omron Corp Start control device for movable body

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2014021421A1 (en) * 2012-08-03 2014-02-06 クラリオン株式会社 Camera parameter computation device, navigation system, and camera parameter computation method
JP2014203349A (en) * 2013-04-08 2014-10-27 スズキ株式会社 Vehicle drive support device
JP2016192104A (en) * 2015-03-31 2016-11-10 キヤノンマーケティングジャパン株式会社 Information processor, processing method for the same, and program

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112241677A (en) * 2019-07-17 2021-01-19 本田技研工业株式会社 Information providing device, information providing method, and storage medium
US20220126818A1 (en) * 2020-10-28 2022-04-28 Toyota Research Institute, Inc. Systems and methods for identifying high-risk driving situations from driving data
US12097845B2 (en) * 2020-10-28 2024-09-24 Toyota Research Institute, Inc. Systems and methods for identifying high-risk driving situations from driving data

Also Published As

Publication number Publication date
JPWO2018127962A1 (en) 2019-06-27
JP6661796B2 (en) 2020-03-11

Similar Documents

Publication Publication Date Title
KR102070530B1 (en) Operation method and system of autonomous vehicle based on motion plan
US11040726B2 (en) Alarm system of autonomous driving vehicles (ADVs)
EP3324332B1 (en) Method and system to predict vehicle traffic behavior for autonomous vehicles to make driving decisions
US10816973B2 (en) Utilizing rule-based and model-based decision systems for autonomous driving control
US20190071101A1 (en) Driving assistance method, driving assistance device which utilizes same, autonomous driving control device, vehicle, driving assistance system, and program
EP3315388A1 (en) Spring system-based change lane approach for autonomous vehicles
JP6412070B2 (en) Driving support device and driving support method
JP6552992B2 (en) Information processing apparatus, in-vehicle apparatus, and information processing method
KR20190013688A (en) Traffic forecasting based on map images for autonomous driving
JP6305650B2 (en) Automatic driving device and automatic driving method
US11508161B2 (en) Driving support system and server device
US11092458B2 (en) Navigation system with operation obstacle alert mechanism and method of operation thereof
JP2020147148A (en) Information processing apparatus and automatic travel control system including information processing apparatus
WO2018127962A1 (en) Reporting control device and reporting control method
WO2019131388A1 (en) Drive assistance device, drive assistance system, drive assistance method, and recording medium in which drive assistance program is stored
CN114802251A (en) Control method and device for automatic driving vehicle, electronic device and storage medium
CN113170092B (en) Image processing apparatus, image processing method, and image processing system
US10543852B2 (en) Environmental driver comfort feedback for autonomous vehicle
JP2020126356A (en) Information processing device, information processing method, and program
US12017585B2 (en) Abnormality detection device, abnormality detection method, program, and information processing system
JP2019016227A (en) Information processing method, information processing apparatus, and information processing program
JP7047001B2 (en) Traffic risk reduction programs, information processing equipment and methods
US20210158692A1 (en) Information processing device, information processing system, and computer readable recording medium
JP2018169945A (en) Driving support apparatus, driving support method, and driving support program
CN112829757A (en) Automatic driving control system, server device, and storage medium for storing programs

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17889571

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2018560292

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17889571

Country of ref document: EP

Kind code of ref document: A1