WO2005076661A1 - 超指向性スピーカ搭載型移動体 - Google Patents
超指向性スピーカ搭載型移動体 Download PDFInfo
- Publication number
- WO2005076661A1 WO2005076661A1 PCT/JP2005/002044 JP2005002044W WO2005076661A1 WO 2005076661 A1 WO2005076661 A1 WO 2005076661A1 JP 2005002044 W JP2005002044 W JP 2005002044W WO 2005076661 A1 WO2005076661 A1 WO 2005076661A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- module
- speaker
- super
- sound
- visual
- Prior art date
Links
- 230000000007 visual effect Effects 0.000 claims abstract description 35
- 230000009471 action Effects 0.000 abstract description 6
- 238000010586 diagram Methods 0.000 description 13
- 230000001360 synchronised effect Effects 0.000 description 8
- 230000004807 localization Effects 0.000 description 6
- 230000015572 biosynthetic process Effects 0.000 description 5
- 238000001514 detection method Methods 0.000 description 5
- 239000000284 extract Substances 0.000 description 5
- 238000003786 synthesis reaction Methods 0.000 description 5
- 238000006243 chemical reaction Methods 0.000 description 3
- 238000000605 extraction Methods 0.000 description 3
- 238000000034 method Methods 0.000 description 3
- 230000005540 biological transmission Effects 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 230000033001 locomotion Effects 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 230000005236 sound signal Effects 0.000 description 2
- 241000282412 Homo Species 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 239000012141 concentrate Substances 0.000 description 1
- 230000001276 controlling effect Effects 0.000 description 1
- 230000002596 correlated effect Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 239000011521 glass Substances 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000001902 propagating effect Effects 0.000 description 1
- 230000035945 sensitivity Effects 0.000 description 1
- 238000001228 spectrum Methods 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R3/00—Circuits for transducers, loudspeakers or microphones
- H04R3/12—Circuits for transducers, loudspeakers or microphones for distributing signals to two or more loudspeakers
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/20—Arrangements for obtaining desired frequency or directional characteristics
- H04R1/32—Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only
- H04R1/323—Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only for loudspeakers
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2201/00—Details of transducers, loudspeakers or microphones covered by H04R1/00 but not provided for in any of its subgroups
- H04R2201/40—Details of arrangements for obtaining desired directional characteristic by combining a number of identical transducers covered by H04R1/40 but not provided for in any of its subgroups
- H04R2201/401—2D or 3D arrays of transducers
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2217/00—Details of magnetostrictive, piezoelectric, or electrostrictive transducers covered by H04R15/00 or H04R17/00 but not provided for in any of their subgroups
- H04R2217/03—Parametric transducers where sound is generated or captured by the acoustic demodulation of amplitude modulated ultrasonic waves
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2430/00—Signal processing covered by H04R, not provided for in its groups
- H04R2430/20—Processing of the output signals of the acoustic transducers of an array for obtaining a desired directivity characteristic
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R27/00—Public address systems
Definitions
- the present invention relates to a mobile-body-mounted acoustic apparatus that has a super-directional speed force for directionally emitting audible sound to a mobile body having a person tracking function.
- a super-directional speaker uses the principle of a parametric speaker that obtains sound in the audible band using the distortion component generated in the process of propagation of strong ultrasonic waves in the air, and concentrates and propagates the sound in front of it. As a result, it is possible to provide sound with narrow directivity.
- a parametric speaker as disclosed in Patent Document 1.
- Patent Document 2 discloses a robot equipped with an audiovisual system. This mobile auditory vision system enables real-time processing to track vision and hearing for the target, and integrates sensor information such as vision, hearing, motor, etc., and if any information is missing, Also continued pursuit by complementing each other.
- Patent Document 1 Japanese Patent Application Laid-Open No. 2001-346288
- Patent Document 2 Japanese Patent Application Laid-Open No. 2002-264058
- a parametric speaker has a strong directivity as a super-directional speaker, so it is possible to limit the audible area.
- the parametric speaker recognizes a specific listener, and is limited to that listener. I could't send voice.
- the present invention has been made to solve the above-described problem, and has a super-directional speaker mounted on a moving body, so that a specific sound can be transmitted to a specific listening.
- the purpose is to provide. Disclosure of the invention
- a mobile object equipped with a super-directional speaker according to the present invention includes an omnidirectional speaker, a super-directional speaker, and a visual module, a hearing module, a motor control module, and an integrated module that integrates them. By combining them, it is possible to transmit sound to specified and unspecified objects simultaneously.
- omnidirectional speakers By combining omnidirectional speakers, it is possible to transmit sound according to the situation. In other words, selecting a speaker such as a super-directional speaker for private information and an omnidirectional speaker for general information expands the range of information transmission methods. Furthermore, by using multiple super-directional speakers, individual information can be conveyed to individual persons without mixing (crosstalk) with individual sounds.
- FIG. 1 is a front view of a moving body according to the first embodiment.
- FIG. 2 is a side view of the moving body according to the first embodiment.
- FIG. 3 is a diagram showing a sound transmission range of a super-directional speaker and an omnidirectional speaker according to Embodiment 1 of the present invention.
- FIG. 4 is a configuration diagram of a superdirective speaker according to Embodiment 1 of the present invention.
- FIG. 5 is an overall system diagram of the first embodiment.
- FIG. 6 is a diagram showing details of a hearing module of the first embodiment.
- FIG. 7 is a diagram showing details of a visual module according to the first embodiment.
- FIG. 8 is a diagram showing details of a motor control module according to the first embodiment.
- FIG. 9 is a diagram showing details of a dialogue module according to the first embodiment.
- FIG. 10 is a diagram showing details of an integrated module according to the first embodiment.
- FIG. 11 is a diagram showing an area where the camera according to the first embodiment detects an object.
- FIG. 12 is a diagram illustrating an object tracking system according to the first embodiment of the present invention.
- FIG. 13 is a view showing a modification of the first embodiment of the present invention.
- FIG. 14 is a diagram showing another modification of the first embodiment of the present invention.
- FIG. 15 is a diagram when the moving object according to the first embodiment of the present invention measures a distance to an object.
- FIG. 1 is a front view of the moving body according to the first embodiment
- FIG. 2 is a side view of the moving body according to the first embodiment.
- a moving object 1 which is a robot having a humanoid appearance includes a leg 2, a torso 3 supported on the leg 2, and a head movably supported on the torso 3.
- the leg 2 is provided with a plurality of wheels 21 at a lower portion, and is movable by controlling a motor described later. Further, the moving mode may include a plurality of leg moving means that are connected only by wheels.
- the body 3 is fixedly supported on the leg 2.
- the head 4 is connected to the body 3 via a connecting member 5, and the connecting member 5 is rotatably supported on the body 3 with respect to a vertical axis as shown by an arrow A.
- the head 4 is supported by the connecting member 5 so as to be rotatable in the vertical direction as shown by the arrow B.
- the head 4 is entirely covered with a soundproof exterior 41, and has a camera 42 as a visual device in charge of robot vision on the front side, and a robot auditory device on both sides for robot hearing.
- a pair of microphones 43 is provided as a device.
- the microphones 43 are mounted on the side surfaces of the head 4 such that the microphones 43 are directed forward and have directivity.
- the omnidirectional speaker 31 is provided on the front surface of the body 3, and the head 4 has a radiating portion, which is a radiating portion of a super directional speaker having high directivity based on the principle of a norametric speed power array.
- a vessel 44 is provided.
- a parametric speaker uses ultrasonic waves that cannot be heard by humans, generates a distortion component in the process of propagation of strong ultrasonic waves in the air, and obtains sound in the audible band by using the distortion component.
- the principle non-linearity
- the conversion efficiency for obtaining audible sound is low, the “super directivity” in which sound is concentrated in a beam in a narrow area in the direction of sound emission is considered.
- Omni-directional speakers form a sound field in a large area including the back surface, like the light of a bare light bulb, so it was impossible to control the area. It also makes it possible to limit the area that sounds like a spotlight.
- FIG. 3 shows the sound propagation between the omnidirectional speaker and the super-directional speaker.
- the upper part of Fig. 3 is a contour diagram of the sound pressure level of the sound propagating in the air, and the lower part is a figure showing the measured values of the sound pressure level.
- Fig. 3 (a) it is clear that the omnidirectional speaker spreads out and can be heard in the surrounding space.
- the sound of the super-directional speaker propagates intensively in front. This utilizes the principle of a parametric loudspeaker that obtains sound in the audible band by using the distortion component generated during the propagation of powerful ultrasonic waves through the air.
- FIG. 3 (b) it is possible to provide sound with narrow directivity.
- the super-directional speaker system includes a sound source 32 from an audible sound signal source and a modulator that modulates an ultrasonic carrier signal with an input electric signal from a signal from the sound source 32. 33, a power amplifier 34 for amplifying a signal from the modulator 33, and a radiator 44 for converting a signal obtained by the modulation into a sound wave.
- a modulator that extracts an audio signal and emits an ultrasonic wave according to the magnitude of the signal is required. Since it can be extracted in a simple manner and fine adjustment can be easily performed, it is more preferable to use an envelope modulator that performs digital processing.
- FIG. 5 shows an electrical configuration of the control system of the moving object.
- the control system includes a network 100, a visual module 300, a visual module 200, a motor control module 400, a dialog module 500, and an integrated module 600.
- the hearing module 300, the vision module 200, the motor control module 400, the dialogue module 500, and the integrated module 600 will be described.
- FIG. 6 shows a detailed view of the hearing module.
- the auditory module 300 includes a microphone 43, a peak detector 301, a sound source localization unit 302, and an auditory event generator 304.
- the hearing module 300 based on an acoustic signal from the microphone 43, Extracts a series of peaks for each of the left and right channels, and pairs the same or similar peaks in the left and right channels.
- the peak extraction is performed by using a band-pass filter that passes only data under the condition that the power is equal to or higher than the threshold value and has a maximum value, for example, a frequency between 90 Hz and 3 kHz.
- This threshold is defined as the value obtained by measuring the background noise in the surroundings and adding a sensitivity parameter, for example, 10 dB.
- the hearing module 300 uses the fact that each peak has a harmonic structure, finds a more accurate peak between the left and right channels, and extracts a sound having a harmonic structure.
- the peak detector 301 analyzes the frequency of the sound input from the microphone 43, detects a peak from the obtained spectrum, and extracts a peak having a harmonic structure from the obtained peak.
- the sound source localization unit 302 localizes the sound source direction in the robot coordinate system by selecting an acoustic signal having the same peak frequency from the left and right channels for each of the extracted peaks, and obtaining a binaural phase difference.
- the auditory event generation unit 304 generates an auditory event 305 including the sound source direction localized by the sound source localization unit 302 and the localization time, and outputs the event to the network 100. When a plurality of harmonic structures are extracted by the peak detection unit 301, a plurality of auditory events 305 are output.
- FIG. 7 shows a detailed view of the visual module.
- the visual module 200 comprises a camera 42, a face finding section 201, a face identifying section 202, a face localizing section 203, a visual event generating section 206, and a face database 208!
- the visual module 200 extracts a face image area of each speaker by, for example, skin color extraction by the face detection unit 201 based on the captured image from the camera, and is registered in the face database 208 in advance by the face identification unit 202.
- the face ID 204 is determined and identified as the face, and the position of the face image area extracted by the face localization unit 203 on the captured image is determined.
- the face position 205 in the robot coordinate system is determined from the size.
- the visual event generation unit 206 generates a visual event 210 including the face ID 204, the face position 205, and the time when these were detected, and outputs the visual event 210 to the network.
- the face recognition unit 202 performs a database search on the extracted face image region using, for example, template matching which is a known image processing described in Patent Document 1.
- the face database 208 It is a database in which each person's face image and name correspond one-to-one and IDs are assigned.
- the visual module 200 performs the above-described processing, that is, identification and localization, on each face.
- the face detection unit 201 performs face area detection, and performs pattern matching based on skin color extraction and correlation calculation. The combination of allows multiple faces to be detected accurately!
- FIG. 8 shows a detailed view of the motor control module.
- the motor control module 400 includes a motor 401 and a potentiometer 402, a PWM control circuit 403, an AD conversion circuit 404, a motor control unit 405, a motor event generation unit 407, a wheel 21, a robot It comprises a head 4, a radiator 44, and an omnidirectional speaker 31.
- the motor control module 400 plans the operation of the moving body 1 based on the attention direction 608 obtained from the integrated module 600 described later. If the operation of the drive motor 401 is required, the motor control module 400 405 controls the drive of the motor 401 via the PWM control circuit 403.
- the motion planning is performed, for example, by moving wheels to move the position of the moving body 1 or moving the position of the moving body 1 so as to move toward the target based on the information on the direction of attention.
- the motor that rotates the head 4 in the horizontal direction is controlled so that the head 4 is directed toward the target.
- the radiator 44 is directed to the position of the head of the object, such as when V The motor that rotates the up and down 4 is controlled, and the direction of the radiator 44 is controlled.
- the motor control module 400 controls the driving of the motor 401 via the PWM control circuit 403, detects the rotation direction of the motor with the potentiometer 402, and detects the direction of the moving body by the motor control unit 405 via the AD conversion circuit 404. 406 is extracted, a motor event generation unit 407 generates a motor event 409 including motor direction information and time force, and outputs the motor event 409 to the network 100.
- FIG. 9 shows a detailed view of the dialogue module.
- the dialogue module 500 includes a speaker, a speech synthesis circuit 501, a dialogue control circuit 502, and a dialogue scenario 503.
- the dialogue module 500 controls the dialogue control circuit 502 based on the face ID 204 obtained by the integrated module 600 described later and the dialogue scenario 503, and drives the omnidirectional speaker 31 by the voice synthesis circuit 501, Output the sound of
- the speech synthesis circuit 501 also functions as a sound source of a super-directional speaker with a highly directional parametric action, and outputs a predetermined sound to a target speaker.
- the dialogue scenario 503 describes to whom and what to speak at what kind of timing, and the dialogue control circuit 502 incorporates the name included in the face ID 204 into the dialogue scenario 503, which is described in the dialogue scenario 503. According to the timing, the content described in the dialogue scenario 503 is synthesized by the voice synthesis circuit 501, and the super directional speaker or the omnidirectional speaker 31 is driven. Further, switching and use of the omnidirectional type force 31 and the radiator 44 are controlled by the dialogue control circuit 502.
- the radiator 44 is configured to transmit sound to a specific listener and a specific area in synchronization with the object tracking means, and the omnidirectional speaker 31 is configured to transmit shared information to an unspecified large number of objects. I have.
- the object can be tracked using the hearing module, the motor control module, the integrated module, and the network among the above configurations (object tracking means). Furthermore, tracking accuracy can be improved by adjusting the visual module.
- the direction of the radiator 44 can be controlled by using the integrated module, the motor control module, the dialog module, and the network (radiator direction control means).
- FIG. 10 shows a detailed view of the integrated module.
- the integration module 600 integrates the auditory module 300, the vision module 200, and the motor control module 400 described above, and generates an input of the interaction module 500. More specifically, the integrated module 600 includes a synchronization circuit 602 that synchronizes the asynchronous event 601a, that is, the auditory event 305, the visual event 210, and the motor event 409 from the auditory module 300, the visual module 200, and the motor control module 400 into a synchronous event 601b. And a stream generation unit 603 for associating the synchronization events 601b with each other to generate an auditory stream 605, a visual stream 606, and an integrated stream 607, and an attention control module 604.
- a synchronization circuit 602 that synchronizes the asynchronous event 601a, that is, the auditory event 305, the visual event 210, and the motor event 409 from the auditory module 300, the visual module 200, and the motor control module 400 into a synchronous event 60
- Synchronization circuit 602 generates auditory event 305 from auditory module 300, visual event 210 with visual module 200 power, and motor event 409 from motor control module 400. Synchronize to generate synchronous auditory events, synchronous visual events, and synchronous motor events. At that time, the synchronous auditory event and the synchronous visual event are converted into an absolute coordinate system using the synchronous motor event.
- the synchronized events are connected in the time direction, and an auditory event forms an auditory stream, and a visual event forms a visual stream.
- an auditory event forms an auditory stream
- a visual event forms a visual stream.
- a plurality of sounds and faces are present at the same time, a plurality of auditory and visual streams are formed.
- the highly correlated visual and auditory streams are bundled together (association) to form an integrated stream t, a higher-order stream.
- the attention control module refers to the sound source direction information included in the formed auditory, visual, and integrated streams to determine the direction 608 to which attention is directed.
- the priority order of the stream reference is the integrated stream, the auditory stream, and then the visual stream. If there is an integrated stream, the sound source direction of the integrated stream, if there is no integrated stream, the audio stream, the integrated stream and the audio stream If there is no audio stream, the direction of the sound source of the visual stream is set to the direction of attention 608.
- Information about the place to be used is input to the moving object in advance, and it is set in advance in which position in the room which directional sound should be heard and how to move. If a human cannot see from the sound source direction due to an obstacle such as a wall, the moving object determines that the human is hidden, and sets the object tracking means in advance so that it takes an action (movement) to search for a face. Keep it.
- the camera 42 of the moving body 1 is provided in front of the head 4, and a projecting range 49 is limited to a part in front of the camera 42 as shown in FIG. For example, if the room has an obstacle E as shown in Fig. 12, it may not be possible to detect visitors.
- the moving body 1 when the moving body 1 is at the position A and the sound source direction is B, if the visitor C cannot be found, the moving body 1 is controlled by the motor control module 800 so as to force in the direction of D. Good. It is set so that blind spots in the field of view due to obstacles E and the like can be eliminated by such active actions. In addition, by using the reflection, it is possible for the mobile unit 1 to transmit the voice to the visitor C without taking the action of D.
- the object tracking means can integrate the auditory information and the visual information and robustly perceive the surrounding situation. It also integrates audiovisual processing and actions By perceiving the surrounding situation more robustly, the scene analysis can be improved.
- the mobile unit 1 waiting in the room controls the wheels 21 and the motors for moving the head so that the camera of the mobile unit faces in a direction in which sound is generated when a person enters the room. Control.
- the dialogue module 500 identifies the name based on the face ID obtained from the integrated module, and uses speech synthesis from the omnidirectional loudspeaker 31 or the radiator 44, which is the radiating part of the super-directional loudspeaker. , And say hello to the visitors.
- the dialogue module 500 controls the dialogue control circuit, and the omnidirectional speaker 31 emits a synthesized voice so that everyone can hear "Welcome everyone.” Judge each person using the visual module 200 as if there were only one visitor.
- radiator 44 which is a super-directional speaker, is used, it cannot be heard by other people, so only the queried visitor answers his or her name. Visitors can be registered.
- the use of a super-directional speaker allows information to be transmitted only to specific visitors.
- An object tracking means composed of an object tracking system that recognizes and tracks an object, and an object tracking system that controls the radiator to face the object tracked by the object tracking means With the direction control means, sound can be transmitted only to a specific target.
- radiator 44 and the camera 42 which are radiating portions of the super-directional speaker, are installed on the head 4 has been described.
- Speedy If the directions of the radiator 44 and the camera 42, which are the power radiating parts, are made variable, the setting place of the radiating part 44 and the power camera 42 is not limited to the head 4 but may be V or offset! ,.
- radiator 44 may be provided so that the directions of the radiators 44 can be individually controlled. It will be possible to convey separate voices only to specific people.
- the video from the camera 42 may be subjected to image processing, and individual sounds may be transmitted from the radiator 44 to a group having a characteristic such as a person wearing glasses. Also, if there are foreigners in the group, the same may be conveyed in English or French, in the native language of the person.
- the mobile object equipped with a super-directional speaker has an omnidirectional speaker and a super-directional speaker, and integrates a visual module, a hearing module, and a motor control module.
- a visual module By having an integrated module, it can transmit sound to specified and unspecified objects at the same time, and is suitable for use in robots equipped with audiovisual systems.
Landscapes
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Otolaryngology (AREA)
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Acoustics & Sound (AREA)
- Signal Processing (AREA)
- Manipulator (AREA)
- Circuit For Audible Band Transducer (AREA)
- Toys (AREA)
- Obtaining Desirable Characteristics In Audible-Bandwidth Transducers (AREA)
- Control Of Position, Course, Altitude, Or Attitude Of Moving Bodies (AREA)
Abstract
Description
Claims
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2005517825A JPWO2005076661A1 (ja) | 2004-02-10 | 2005-02-10 | 超指向性スピーカ搭載型移動体 |
US10/588,801 US20070183618A1 (en) | 2004-02-10 | 2005-02-10 | Moving object equipped with ultra-directional speaker |
EP05710096A EP1715717B1 (en) | 2004-02-10 | 2005-02-10 | Moving object equipped with ultra-directional speaker |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2004033979 | 2004-02-10 | ||
JP2004-033979 | 2004-02-10 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2005076661A1 true WO2005076661A1 (ja) | 2005-08-18 |
Family
ID=34836159
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2005/002044 WO2005076661A1 (ja) | 2004-02-10 | 2005-02-10 | 超指向性スピーカ搭載型移動体 |
Country Status (4)
Country | Link |
---|---|
US (1) | US20070183618A1 (ja) |
EP (1) | EP1715717B1 (ja) |
JP (1) | JPWO2005076661A1 (ja) |
WO (1) | WO2005076661A1 (ja) |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2009111833A (ja) * | 2007-10-31 | 2009-05-21 | Mitsubishi Electric Corp | 情報提示装置 |
JP2009531926A (ja) * | 2006-03-31 | 2009-09-03 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | データ処理装置及び方法 |
JP2011520496A (ja) * | 2008-05-14 | 2011-07-21 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | 対話形システム及び方法 |
WO2012032704A1 (ja) * | 2010-09-08 | 2012-03-15 | パナソニック株式会社 | 音響再生装置 |
JP2012175162A (ja) * | 2011-02-17 | 2012-09-10 | Waseda Univ | 音響システム |
US9036856B2 (en) | 2013-03-05 | 2015-05-19 | Panasonic Intellectual Property Management Co., Ltd. | Sound reproduction device |
JP2016206646A (ja) * | 2015-04-24 | 2016-12-08 | パナソニックIpマネジメント株式会社 | 音声再生方法、音声対話装置及び音声対話プログラム |
KR20170027804A (ko) * | 2014-06-27 | 2017-03-10 | 마이크로소프트 테크놀로지 라이센싱, 엘엘씨 | 지향성 오디오 통지 |
JP2017170568A (ja) * | 2016-03-24 | 2017-09-28 | 株式会社国際電気通信基礎技術研究所 | サービス提供ロボットシステム |
JP2019501606A (ja) * | 2015-11-04 | 2019-01-17 | ズークス インコーポレイテッド | 音響ビームフォーミングを介した外部環境とのロボット車両通信のための方法 |
WO2020202621A1 (ja) * | 2019-03-29 | 2020-10-08 | パナソニックIpマネジメント株式会社 | 無人移動体及び情報処理方法 |
Families Citing this family (42)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
ATE400474T1 (de) * | 2005-02-23 | 2008-07-15 | Harman Becker Automotive Sys | Spracherkennungssytem in einem kraftfahrzeug |
JP5170961B2 (ja) * | 2006-02-01 | 2013-03-27 | ソニー株式会社 | 画像処理システム、画像処理装置および方法、プログラム、並びに記録媒体 |
JP2007282191A (ja) * | 2006-03-14 | 2007-10-25 | Seiko Epson Corp | 案内装置及びその制御方法 |
WO2009104117A1 (en) * | 2008-02-18 | 2009-08-27 | Koninklijke Philips Electronics N.V. | Light controlled audio transducer |
JPWO2010041394A1 (ja) * | 2008-10-06 | 2012-03-01 | パナソニック株式会社 | 音響再生装置 |
KR20100119342A (ko) * | 2009-04-30 | 2010-11-09 | 삼성전자주식회사 | 디스플레이장치 및 그 제어방법 |
US8515092B2 (en) * | 2009-12-18 | 2013-08-20 | Mattel, Inc. | Interactive toy for audio output |
DE202009017384U1 (de) * | 2009-12-22 | 2010-03-25 | Metallbau & Schweißtechnologie Zentrum GmbH Blankenburg | Automat zur Erstellung von individuellen, standortbezogenen Bildern, Grußkarten u.dgl. |
TWI394143B (zh) * | 2010-07-30 | 2013-04-21 | Hwa Hsia Inst Of Technology | 機器人視覺及聽覺的阻絕裝置 |
WO2012060041A1 (ja) * | 2010-11-01 | 2012-05-10 | 日本電気株式会社 | 発振装置及び携帯装置 |
WO2013012412A1 (en) * | 2011-07-18 | 2013-01-24 | Hewlett-Packard Development Company, L.P. | Transmit audio in a target space |
US8666107B2 (en) * | 2012-04-11 | 2014-03-04 | Cheng Uei Precision Industry Co., Ltd. | Loudspeaker |
KR101428877B1 (ko) * | 2012-12-05 | 2014-08-14 | 엘지전자 주식회사 | 로봇 청소기 |
US10575093B2 (en) | 2013-03-15 | 2020-02-25 | Elwha Llc | Portable electronic device directed audio emitter arrangement system and method |
US10181314B2 (en) * | 2013-03-15 | 2019-01-15 | Elwha Llc | Portable electronic device directed audio targeted multiple user system and method |
US9886941B2 (en) | 2013-03-15 | 2018-02-06 | Elwha Llc | Portable electronic device directed audio targeted user system and method |
US10291983B2 (en) * | 2013-03-15 | 2019-05-14 | Elwha Llc | Portable electronic device directed audio system and method |
US10531190B2 (en) | 2013-03-15 | 2020-01-07 | Elwha Llc | Portable electronic device directed audio system and method |
CN104065798B (zh) * | 2013-03-21 | 2016-08-03 | 华为技术有限公司 | 声音信号处理方法及设备 |
US9560449B2 (en) | 2014-01-17 | 2017-01-31 | Sony Corporation | Distributed wireless speaker system |
US9866986B2 (en) | 2014-01-24 | 2018-01-09 | Sony Corporation | Audio speaker system with virtual music performance |
US9232335B2 (en) | 2014-03-06 | 2016-01-05 | Sony Corporation | Networked speaker system with follow me |
HK1195445A2 (en) * | 2014-05-08 | 2014-11-07 | 黃偉明 | Endpoint mixing system and reproduction method of endpoint mixed sounds |
US9544679B2 (en) | 2014-12-08 | 2017-01-10 | Harman International Industries, Inc. | Adjusting speakers using facial recognition |
US9693168B1 (en) | 2016-02-08 | 2017-06-27 | Sony Corporation | Ultrasonic speaker assembly for audio spatial effect |
US9826332B2 (en) | 2016-02-09 | 2017-11-21 | Sony Corporation | Centralized wireless speaker system |
US9924291B2 (en) | 2016-02-16 | 2018-03-20 | Sony Corporation | Distributed wireless speaker system |
US9826330B2 (en) | 2016-03-14 | 2017-11-21 | Sony Corporation | Gimbal-mounted linear ultrasonic speaker assembly |
US9693169B1 (en) | 2016-03-16 | 2017-06-27 | Sony Corporation | Ultrasonic speaker assembly with ultrasonic room mapping |
US9794724B1 (en) | 2016-07-20 | 2017-10-17 | Sony Corporation | Ultrasonic speaker assembly using variable carrier frequency to establish third dimension sound locating |
US9924286B1 (en) | 2016-10-20 | 2018-03-20 | Sony Corporation | Networked speaker system with LED-based wireless communication and personal identifier |
US9854362B1 (en) | 2016-10-20 | 2017-12-26 | Sony Corporation | Networked speaker system with LED-based wireless communication and object detection |
US10075791B2 (en) | 2016-10-20 | 2018-09-11 | Sony Corporation | Networked speaker system with LED-based wireless communication and room mapping |
CN107105369A (zh) * | 2017-06-29 | 2017-08-29 | 京东方科技集团股份有限公司 | 声音定向切换装置及显示系统 |
EP3696811A4 (en) * | 2017-10-11 | 2020-11-25 | Sony Corporation | VOICE ENTRY DEVICE, ASSOCIATED PROCESS AND PROGRAM |
CN107864430A (zh) * | 2017-11-03 | 2018-03-30 | 杭州聚声科技有限公司 | 一种声波定向传播控制系统及其控制方法 |
CN108931979B (zh) * | 2018-06-22 | 2020-12-15 | 中国矿业大学 | 基于超声波辅助定位的视觉跟踪移动机器人及控制方法 |
CN109217943A (zh) * | 2018-07-19 | 2019-01-15 | 珠海格力电器股份有限公司 | 定向播报方法、装置、家用电器及计算机可读存储介质 |
US10623859B1 (en) | 2018-10-23 | 2020-04-14 | Sony Corporation | Networked speaker system with combined power over Ethernet and audio delivery |
US11140477B2 (en) | 2019-01-06 | 2021-10-05 | Frank Joseph Pompei | Private personal communications device |
US11443737B2 (en) | 2020-01-14 | 2022-09-13 | Sony Corporation | Audio video translation into multiple languages for respective listeners |
US11256878B1 (en) | 2020-12-04 | 2022-02-22 | Zaps Labs, Inc. | Directed sound transmission systems and methods |
Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH02230898A (ja) | 1989-03-03 | 1990-09-13 | Nippon Telegr & Teleph Corp <Ntt> | 音声再生方式 |
JPH11258101A (ja) * | 1998-03-13 | 1999-09-24 | Honda Motor Co Ltd | 自動車のリーク検査装置 |
JP2001346288A (ja) | 2000-06-02 | 2001-12-14 | Mk Seiko Co Ltd | パラメトリックスピーカー |
JP2002264058A (ja) | 2001-03-09 | 2002-09-18 | Japan Science & Technology Corp | ロボット視聴覚システム |
JP2003251583A (ja) * | 2002-03-01 | 2003-09-09 | Japan Science & Technology Corp | ロボット視聴覚システム |
JP2003285286A (ja) * | 2002-03-27 | 2003-10-07 | Nec Corp | ロボット装置 |
JP2003340764A (ja) * | 2002-05-27 | 2003-12-02 | Matsushita Electric Works Ltd | 案内ロボット |
EP1375084A1 (en) | 2001-03-09 | 2004-01-02 | Japan Science and Technology Corporation | Robot audiovisual system |
JP2004286805A (ja) * | 2003-03-19 | 2004-10-14 | Sony Corp | 話者識別装置および話者識別方法、並びにプログラム |
JP2004295059A (ja) * | 2003-03-27 | 2004-10-21 | Katsuyoshi Mizuno | 映像情報及び音声情報を連動した画像が平面移動する方法。 |
WO2004093488A2 (en) | 2003-04-15 | 2004-10-28 | Ipventure, Inc. | Directional speakers |
JP2004318026A (ja) * | 2003-04-14 | 2004-11-11 | Tomohito Nakagawa | セキュリティペットロボット及びその装置に関する信号処理方法 |
Family Cites Families (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5796819A (en) * | 1996-07-24 | 1998-08-18 | Ericsson Inc. | Echo canceller for non-linear circuits |
US6914622B1 (en) * | 1997-05-07 | 2005-07-05 | Telbotics Inc. | Teleconferencing robot with swiveling video monitor |
JP4221792B2 (ja) * | 1998-01-09 | 2009-02-12 | ソニー株式会社 | スピーカ装置及びオーディオ信号送信装置 |
JP2000023281A (ja) * | 1998-04-28 | 2000-01-21 | Canon Inc | 音声出力装置および方法 |
DE19935375C1 (de) * | 1999-07-29 | 2001-07-05 | Bosch Gmbh Robert | Verfahren und Vorrichtung zur geräuschabhängigen Ansteuerung von Aggregaten in einem Fahrzeug |
US6894714B2 (en) * | 2000-12-05 | 2005-05-17 | Koninklijke Philips Electronics N.V. | Method and apparatus for predicting events in video conferencing and other applications |
JP2003023689A (ja) * | 2001-07-09 | 2003-01-24 | Sony Corp | 可変指向性超音波スピーカシステム |
WO2003019125A1 (en) * | 2001-08-31 | 2003-03-06 | Nanyang Techonological University | Steering of directional sound beams |
US20030063756A1 (en) * | 2001-09-28 | 2003-04-03 | Johnson Controls Technology Company | Vehicle communication system |
US6690802B2 (en) * | 2001-10-24 | 2004-02-10 | Bestop, Inc. | Adjustable speaker box for the sports bar of a vehicle |
US7139401B2 (en) * | 2002-01-03 | 2006-11-21 | Hitachi Global Storage Technologies B.V. | Hard disk drive with self-contained active acoustic noise reduction |
JP3902551B2 (ja) * | 2002-05-17 | 2007-04-11 | 日本ビクター株式会社 | 移動ロボット |
US20040114770A1 (en) * | 2002-10-30 | 2004-06-17 | Pompei Frank Joseph | Directed acoustic sound system |
US7983920B2 (en) * | 2003-11-18 | 2011-07-19 | Microsoft Corporation | Adaptive computing environment |
US7492913B2 (en) * | 2003-12-16 | 2009-02-17 | Intel Corporation | Location aware directed audio |
JP4349123B2 (ja) * | 2003-12-25 | 2009-10-21 | ヤマハ株式会社 | 音声出力装置 |
-
2005
- 2005-02-10 US US10/588,801 patent/US20070183618A1/en not_active Abandoned
- 2005-02-10 WO PCT/JP2005/002044 patent/WO2005076661A1/ja not_active Application Discontinuation
- 2005-02-10 EP EP05710096A patent/EP1715717B1/en not_active Not-in-force
- 2005-02-10 JP JP2005517825A patent/JPWO2005076661A1/ja active Pending
Patent Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH02230898A (ja) | 1989-03-03 | 1990-09-13 | Nippon Telegr & Teleph Corp <Ntt> | 音声再生方式 |
JPH11258101A (ja) * | 1998-03-13 | 1999-09-24 | Honda Motor Co Ltd | 自動車のリーク検査装置 |
JP2001346288A (ja) | 2000-06-02 | 2001-12-14 | Mk Seiko Co Ltd | パラメトリックスピーカー |
JP2002264058A (ja) | 2001-03-09 | 2002-09-18 | Japan Science & Technology Corp | ロボット視聴覚システム |
EP1375084A1 (en) | 2001-03-09 | 2004-01-02 | Japan Science and Technology Corporation | Robot audiovisual system |
JP2003251583A (ja) * | 2002-03-01 | 2003-09-09 | Japan Science & Technology Corp | ロボット視聴覚システム |
JP2003285286A (ja) * | 2002-03-27 | 2003-10-07 | Nec Corp | ロボット装置 |
JP2003340764A (ja) * | 2002-05-27 | 2003-12-02 | Matsushita Electric Works Ltd | 案内ロボット |
JP2004286805A (ja) * | 2003-03-19 | 2004-10-14 | Sony Corp | 話者識別装置および話者識別方法、並びにプログラム |
JP2004295059A (ja) * | 2003-03-27 | 2004-10-21 | Katsuyoshi Mizuno | 映像情報及び音声情報を連動した画像が平面移動する方法。 |
JP2004318026A (ja) * | 2003-04-14 | 2004-11-11 | Tomohito Nakagawa | セキュリティペットロボット及びその装置に関する信号処理方法 |
WO2004093488A2 (en) | 2003-04-15 | 2004-10-28 | Ipventure, Inc. | Directional speakers |
Cited By (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2009531926A (ja) * | 2006-03-31 | 2009-09-03 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | データ処理装置及び方法 |
JP2009111833A (ja) * | 2007-10-31 | 2009-05-21 | Mitsubishi Electric Corp | 情報提示装置 |
JP2011520496A (ja) * | 2008-05-14 | 2011-07-21 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | 対話形システム及び方法 |
WO2012032704A1 (ja) * | 2010-09-08 | 2012-03-15 | パナソニック株式会社 | 音響再生装置 |
JP5212575B2 (ja) * | 2010-09-08 | 2013-06-19 | パナソニック株式会社 | 音響再生装置 |
US8750543B2 (en) | 2010-09-08 | 2014-06-10 | Panasonic Corporation | Sound reproduction device |
US9743186B2 (en) | 2010-09-08 | 2017-08-22 | Panasonic Intellectual Property Management Co., Ltd. | Sound reproduction device |
JP2012175162A (ja) * | 2011-02-17 | 2012-09-10 | Waseda Univ | 音響システム |
US9036856B2 (en) | 2013-03-05 | 2015-05-19 | Panasonic Intellectual Property Management Co., Ltd. | Sound reproduction device |
KR102369879B1 (ko) * | 2014-06-27 | 2022-03-02 | 마이크로소프트 테크놀로지 라이센싱, 엘엘씨 | 지향성 오디오 통지 |
KR20170027804A (ko) * | 2014-06-27 | 2017-03-10 | 마이크로소프트 테크놀로지 라이센싱, 엘엘씨 | 지향성 오디오 통지 |
JP2017525260A (ja) * | 2014-06-27 | 2017-08-31 | マイクロソフト テクノロジー ライセンシング,エルエルシー | 指向性オーディオ通知 |
JP2016206646A (ja) * | 2015-04-24 | 2016-12-08 | パナソニックIpマネジメント株式会社 | 音声再生方法、音声対話装置及び音声対話プログラム |
JP2019501606A (ja) * | 2015-11-04 | 2019-01-17 | ズークス インコーポレイテッド | 音響ビームフォーミングを介した外部環境とのロボット車両通信のための方法 |
US11091092B2 (en) | 2015-11-04 | 2021-08-17 | Zoox, Inc. | Method for robotic vehicle communication with an external environment via acoustic beam forming |
JP2017170568A (ja) * | 2016-03-24 | 2017-09-28 | 株式会社国際電気通信基礎技術研究所 | サービス提供ロボットシステム |
WO2020202621A1 (ja) * | 2019-03-29 | 2020-10-08 | パナソニックIpマネジメント株式会社 | 無人移動体及び情報処理方法 |
JPWO2020202621A1 (ja) * | 2019-03-29 | 2020-10-08 | ||
JP7426631B2 (ja) | 2019-03-29 | 2024-02-02 | パナソニックIpマネジメント株式会社 | 無人移動体及び情報処理方法 |
Also Published As
Publication number | Publication date |
---|---|
EP1715717B1 (en) | 2012-04-18 |
EP1715717A4 (en) | 2009-04-08 |
US20070183618A1 (en) | 2007-08-09 |
JPWO2005076661A1 (ja) | 2008-01-10 |
EP1715717A1 (en) | 2006-10-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2005076661A1 (ja) | 超指向性スピーカ搭載型移動体 | |
US7424118B2 (en) | Moving object equipped with ultra-directional speaker | |
JP3627058B2 (ja) | ロボット視聴覚システム | |
US20090122648A1 (en) | Acoustic mobility aid for the visually impaired | |
US20090030552A1 (en) | Robotics visual and auditory system | |
JP7271695B2 (ja) | ハイブリッドスピーカ及びコンバータ | |
JP2008543143A (ja) | 音響変換器のアセンブリ、システムおよび方法 | |
JP2009514312A (ja) | 音響追跡手段を備える補聴器 | |
JP3632099B2 (ja) | ロボット視聴覚システム | |
EP4358537A2 (en) | Directional sound modification | |
JP2021520760A (ja) | 音源の位置特定 | |
JP2000295698A (ja) | バーチャルサラウンド装置 | |
WO2018086056A1 (zh) | 一种自动识别捕捉人脸定位的组合式音响系统 | |
JP6917107B2 (ja) | 移動体およびプログラム | |
JP3843743B2 (ja) | ロボット視聴覚システム | |
JP3843741B2 (ja) | ロボット視聴覚システム | |
JP2002264058A (ja) | ロボット視聴覚システム | |
JP2002303666A (ja) | マイクロホン装置及び位置検出システム | |
Nakadai et al. | Towards new human-humanoid communication: listening during speaking by using ultrasonic directional speaker | |
Michaud et al. | SmartBelt: A wearable microphone array for sound source localization with haptic feedback | |
US20070041598A1 (en) | System for location-sensitive reproduction of audio signals | |
US20160128891A1 (en) | Method and apparatus for providing space information | |
JP2024504379A (ja) | マイクロホンビームステアリングを用いる頭部装着型コンピューティングデバイス | |
JP2001215989A (ja) | ロボット聴覚システム | |
JP2005176221A (ja) | 音響システム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWE | Wipo information: entry into national phase |
Ref document number: 2005517825 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2005710096 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 10588801 Country of ref document: US Ref document number: 2007183618 Country of ref document: US |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWW | Wipo information: withdrawn in national office |
Country of ref document: DE |
|
WWP | Wipo information: published in national office |
Ref document number: 2005710096 Country of ref document: EP |
|
WWP | Wipo information: published in national office |
Ref document number: 10588801 Country of ref document: US |