WO2000054252A2 - Method with a plurality of speech recognizers - Google Patents
Method with a plurality of speech recognizers Download PDFInfo
- Publication number
- WO2000054252A2 WO2000054252A2 PCT/EP2000/001145 EP0001145W WO0054252A2 WO 2000054252 A2 WO2000054252 A2 WO 2000054252A2 EP 0001145 W EP0001145 W EP 0001145W WO 0054252 A2 WO0054252 A2 WO 0054252A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- speech
- client
- user
- speech input
- recognition
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 30
- 238000004891 communication Methods 0.000 claims abstract description 8
- 238000004458 analytical method Methods 0.000 description 8
- 230000004044 response Effects 0.000 description 8
- 238000000605 extraction Methods 0.000 description 6
- 238000010586 diagram Methods 0.000 description 4
- 239000013598 vector Substances 0.000 description 4
- 230000002457 bidirectional effect Effects 0.000 description 1
- 238000007781 pre-processing Methods 0.000 description 1
- 230000000630 rising effect Effects 0.000 description 1
- 230000005236 sound signal Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/28—Constructional details of speech recognition systems
- G10L15/32—Multiple recognisers used in sequence or in parallel; Score combination systems therefor, e.g. voting systems
Definitions
- the invention relates to a method in which an information unit that makes a speech input possible is stored on a server and can be retrieved by a client.
- EP 0 872 827 describes a system and a method of speech recognition.
- a client on which compressed software for speech recognition is executed is connected to a speech recognition server through a network.
- the client sends a speech recognition grammar and the data of the speech input to the speech recognition server.
- the speech recognition server executes the speech recognition and returns the recognition result to the client.
- the client can be coupled through a communications network to a plurality of speech recognizers and a user's speech input is applied to at least one speech recognizer for the generation of a recognition result and the recognition result is interpreted in a plurality of independent processes and a plurality of interpretation results are generated which are supplied to the user.
- a service provider stores an information unit on a server, which information unit makes a speech input possible.
- a client downloads an information unit from this server, which information unit makes a speech input possible.
- a server is a computer in a communications network, for example, the Internet, on which information of providers is stored and can be retrieved by clients.
- a client is a computer which is connected to a server for retrieving information from the Internet and downloads the information unit stored on the server to represent the information unit by means of software. This information unit is delivered by the client so that the user can perceive the contents of this information unit. The user is requested either by the information unit to enter speech, or, since this information unit has often been invoked, is informed about the possibility of entering speech.
- this speech input is applied to one or more speech recognizers.
- the individual speech recognizers execute a speech recognition and each generate a recognition result.
- These recognition results are each subjected to an interpretation.
- the recognition results are used to come to interpretation results in independent processes.
- this recognition result is analyzed. Therefore, the recognition result is subdivided into its component parts and for example keywords are looked for. Parts of the recognition result that are uninteresting for a later information inquiry are omitted.
- the analysis can then be made from the speech recognizer or from a database. For analyzing the recognition result it is therefore necessary to have information about the contents of the speech input. Possible contents of the speech input are determined by the contents of the information unit.
- Speech recognition for generating recognition results can be used with different cost levels. Speech recognizers are distinguished not only by their size and specialization of the vocabulary, but also by the algorithms with which they perform the speech recognition. A good database inquiry requires a good recognition of this inquiry made by the user via his speech input.
- the interpretation results from the speech recognizer or the database are either automatically sent back to the client, or the server renders them available, so that the user can retrieve the individual interpretation results as required. In either case the interpretation results are delivered by the client in a form that can be perceived by the user.
- the user Due to the combination of the information unit and one or more speech recognizers, the user is provided with a multiple answer to his inquiry made by speech input. As a result, he receives information for which, without this method, he would have to start more than one inquiries with considerable time delay.
- the recognition result is fed to a plurality of interpretation processes which all produce an interpretation result which is sent back to the client or retrieved by him and thus provides a multiple response to the user's inquiry.
- additional software is started on the client when the information unit is loaded, which additional software carries out an extraction of the features of the speech input.
- This additional software digitizes, quantizes and subjects the speech input available as an electric signal to respective analyses, which produce components to which feature vectors are assigned.
- These feature vectors are then transmitted to the coupled speech recognizer.
- the speech recognizer executes the compute- intensive recognition.
- the speech input is compressed and coded, so that the number of data to be transmitted is reduced.
- the time necessary for the feature extraction is reduced on the side of the client, so that the speech recognizer only executes the recognition of the feature vectors applied to it. With speech recognizers that are used frequently, this reduction may be advantageous.
- the speech input is assigned to a plurality of speech recognizers, there is the advantage that the preprocessing needs to be carried out only once. Without the feature extraction on the side of the client, each selected speech recognizer would execute such an extraction.
- the client downloads the information unit in the form of an HTML page (Hyper Text Markup Language) from the server.
- This HTML page is shown by means of a Web browser on the client.
- the client sets up a connection by means of a link to the server, on which link the HTML page, in which the user is interested in, is stored.
- the HTML page can contain graphic symbols, audio and/or video data in addition to text to be represented.
- the HTML page requests the user via an indication to make a speech input. After the user has made this speech input, this speech input is transferred from the client to one or more speech recognizers. A speech recognition is then executed there. The quality of the recognition result then decisively depends on how specialized the speech recognizers are.
- Speech recognizers work with a certain finite vocabulary, which is mostly limited to special fields of application. Therefore, it is important for a usable recognition result that the speech recognizers to which the speech input is transferred are accordingly specialized.
- the recognition result or a plurality of recognition results is/are subjected to an interpretation process. For this purpose, for example the recognized speech input is analyzed for a database and on the basis of this analysis an inquiry is made to the data file of this database.
- the resulting interpretation result is automatically sent back to the client or retrieved by the client and represented there by a Web browser. The user can now make a choice from the plurality of interpretation results. This operation can be compared with looking up in a plurality of lexicons, with the advantage of saving time.
- a speech recognizer connected through the communications network, to which recognizer the speech input coming from the user is sent.
- the speech recognizers execute the speech recognition and convey the individual recognition results to independent interpretation processes.
- the interpretation results sent back to the client or retrieved by him are offered to the user in the form of a graphical representation or as an audio signal.
- the objects which may be realized, for example, as advertising banners are offered by companies working in the same line of business, a user is presented with a plurality of offers from competing firms as a result of his speech input and its multiple parallel processing.
- a user's speech input relating to a specific advertising banner is conveyed to the speech recognizers assigned to an object in that the advertising banner is clicked on with the mouse or in that the user's point of vision is followed, or in that priorities are given to the plurality of speech input options of the individual objects. It is then advantageous to either store the speech input or the preprocessed speech input in a memory on the client, or to send 5 the recognition result back to the client, so that for the purpose of another interpretation process the user can employ this intermediate result which is available anyway.
- the stored speech input or recognition result is then conveyed to another speech recognizer if a speech input has been stored, or to another database if a recognition result has been stored, so as to be capable of making further interpretation results with further interpretations.
- a choice is made from a plurality of objects represented by the Web browser which are enabled by a speech input. From the total number of objects shown, the user chooses several objects, for example, by clicking the mouse. A speech input is then sent only to the speech recognizers of these chosen objects.
- a server assigns additional information in the form of an HTML tag to each object to combine the object with a speech recognizer. As a result, while the HTML page is being downloaded, the object is informed of which speech recognizer on the Internet the speech input is to be sent to be processed.
- a further advantageous embodiment of the invention is provided by the possibility of leaving the decision to which databases the recognition result is sent up to the speech recognizer. This achieves a shift of the decision on which database the user's inquiry is to be processed.
- the HTML page provider who assigns the speech recognizer to the respective object is not up to date as regards the databases, but the operator of the speech recognizers is and he is the one who assigns the databases, the quality of the response to the request is enhanced as a result thereof.
- the HTML page provider who is independent of publishers can send a recognition result from a user's inquiry about new publications in a respective field to all the databases available to him. As a result, the user rapidly receives extensive information about new publications of books of a respective field.
- the object is also achieved by a server on which an information unit is stored which can be retrieved by a client, while there is provided that - the client can be coupled to one or more speech recognizers for generating a plurality of interpretation results sent to a user, and a speech input is applied to at least one speech recognizer for generating recognition results and the recognition results are interpreted in a plurality of independent processes, and for determining a combination of an object that makes a speech input possible with a speech recognizer for generating a recognition result, additional information is assigned to the object.
- Fig. 1 shows a block diagram of an arrangement for implementing the method according to the invention
- Fig. 2 shows a block diagram of the method according to the invention with a speech recognizer
- Fig. 3 shows a block diagram of the method according to the invention with parallel speech recognizers
- Fig. 4 shows a block diagram of the method according to the invention with parallel speech recognizers with an integrated database.
- Fig. 1 shows by way of example an arrangement for implementing the method according to the invention.
- An information unit 3 is stored on a server 1.
- the server 1 can be coupled to a client 2 through a communications network 6.
- This communications network 6, called Internet 6 hereinafter, speech recognizers 7-9 can be coupled to the client 2.
- databases 5 can be coupled to the client 2, to the speech recognizers 7 and 9 and to the server 1.
- a provider stores the information unit 3 on the server 1 to allow a user to access information, for example, via this provider.
- the information unit 3 contains not only contents to be represented and formatting instructions, but also additional information 4.
- the user downloads an information unit 3 which is of interest to him, in the following to be referenced
- HTML page 3 from the server 1.
- a connection based on the TCP/IP protocol is set up to the server 1.
- Software is executed on the client 2, which software may be realized, for example, by a Web browser and by which the HTML page 3 is shown to the user.
- the client 2 includes a memory 25 in which a speech input uttered by the user or a recognition result sent back by a speech recognizer 7-9 is stored.
- Fig. 2 shows the information unit 3 which offers the user interactivity in the form of a speech-input option.
- the objects 19, 20 and 21 are advertising banners, which show the user, for example, advertisements of car firms. Furthermore, they show the user that this HTML page 3 offers a speech input option in that the user, for example, by flashing text - for example, "tell us which car you are interested in" -, utters a speech input. In this example of embodiment all three advertising banners 19, 20, 21 expect to receive a similar speech input. Therefore, the speech input is conveyed to only one speech recognizer 7 via the Internet 6.
- the user can pronounce concepts or word groups of interest to him, which are fed to the client by means of an input device 10 and are conveyed to the speech recognizer 7.
- an extraction of the features of a speech input can be made on the client 2, so that the speech recognizer 7 is only supplied with the speech-input features arranged in feature vectors in compressed form.
- the speech recognizer 7 carries out the speech recognition and generates a recognition result 11.
- This recognition result 11 is analyzed and sent as an inquiry from the speech recognizer 7 to the databases 14, 15 and 16.
- the inquiries, which are in this case sent to the databases 14, 15 and 16, are the same.
- the databases may also be located on the same server as the speech recognizer
- the speech recognizer 7 belongs to the provider of the HTML page 3 or is hired by him. Since the provider knows that inquiries are made after cars on this HTML page 3, the client is connected to a specialized speech recognizer for recognizing the speech input.
- the database 14 contains data from a file of the car firm of advertising banner 19.
- Database 15 contains data of the car firm with advertising banner 20 and the database 16 of the car firm with advertising banner 21.
- the databases 14, 15 and 16 are then searched for information that is in line with the inquiry. This operation is also referenced interpretation.
- the databases 14, 15 and 16 each produce an interpretation result 22, 23 and 24 which is shown on the client 2 after being transmitted via the Internet 6.
- the provider of the HTML page can transfer information that is important for the analysis of a recognition result to the speech recognizers or databases.
- the memory 25 extends the arrangement in that with successive inquiries, the speech input is stored in the memory 25. It is alternatively possible to have this memory 25 store the already generated recognition result. In that case the user can successively inquire a plurality of databases, without repeating each time the speech input or also the speech recognition.
- Fig. 3 shows the arrangement of a method in which a speech input is conveyed to three different speech recognizers 7, 8 and 9.
- the user of the objects 19, 20 and 21 is accordingly requested to utter a speech input.
- This speech utterance is conveyed to the speech recognizers 7, 8 and 9 for generating each a recognition result 11, 12 and 13.
- the speech recognizers 7-9 analyze the recognition results 11, 12 and 13 and prepare each an inquiry for the databases 14, 15 and 16.
- the recognition results 11, 12 and 13 are different, because they were generated by different speech recognizers 7-9 and, on the other hand, different inquiries are generated with these different recognition results 11, 12 and 13 during the analysis, which inquiries are applied to different databases 14, 15 and 16, the user receives with the interpretation results 22, 23 and 24 returned to him on the client 2, three responses based on different databases.
- the databases 14-16 can then make the analyses of the individual recognition results 11, 12 and 13 with key words which are specifically contained in their respective database.
- Fig. 4 shows an arrangement in which the databases 14-16 are integrated with the speech recognizers 7-9. With smaller data files it is possible to integrate the databases 14- 16 with the respective speech recognizers 7-9. Furthermore, there is represented here that a bidirectional link is made from the respective advertising banners 19-21 to the associated interpretation results 22-24 and the associated databases 14-16. It is possible that a response to an inquiry in one of the databases 14-16 is so large that a representation of the interpretation result 22-24 on the client is not wise. In such a case, for example, only the number of found responses to a speech input are sent back to the client and displayed. When the user would like to see the interpretation result 21 of the firm having, for example, advertising banner 19, he can request it and retrieve it from the database 14. These results are then displayed on the client 2.
Landscapes
- Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Information Transfer Between Computers (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Machine Translation (AREA)
- Computer And Data Communications (AREA)
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP00905058A EP1163660A2 (de) | 1999-03-09 | 2000-02-10 | Mehrere spracherkenner verwendendes verfahren |
AU26721/00A AU2672100A (en) | 1999-03-09 | 2000-02-10 | Method with a plurality of speech recognizers |
JP2000604400A JP2002539481A (ja) | 1999-03-09 | 2000-02-10 | 複数の音声認識器を用いる方法 |
KR1020017011408A KR20010108330A (ko) | 1999-03-09 | 2000-02-10 | 복수의 음성 인식자를 구비한 방법 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
DE19910234.1 | 1999-03-09 | ||
DE19910234A DE19910234A1 (de) | 1999-03-09 | 1999-03-09 | Verfahren mit mehreren Spracherkennern |
Publications (2)
Publication Number | Publication Date |
---|---|
WO2000054252A2 true WO2000054252A2 (en) | 2000-09-14 |
WO2000054252A3 WO2000054252A3 (en) | 2000-12-28 |
Family
ID=7900178
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/EP2000/001145 WO2000054252A2 (en) | 1999-03-09 | 2000-02-10 | Method with a plurality of speech recognizers |
Country Status (7)
Country | Link |
---|---|
EP (1) | EP1163660A2 (de) |
JP (1) | JP2002539481A (de) |
KR (1) | KR20010108330A (de) |
CN (1) | CN1350685A (de) |
AU (1) | AU2672100A (de) |
DE (1) | DE19910234A1 (de) |
WO (1) | WO2000054252A2 (de) |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2002103675A1 (en) * | 2001-06-19 | 2002-12-27 | Intel Corporation | Client-server based distributed speech recognition system architecture |
WO2003049080A1 (en) * | 2001-11-30 | 2003-06-12 | Dictaphone Corporation | Distributed speech recognition system with speech recognition engines offering multiple fuctionalites |
US7133829B2 (en) | 2001-10-31 | 2006-11-07 | Dictaphone Corporation | Dynamic insertion of a speech recognition engine within a distributed speech recognition system |
US7146321B2 (en) | 2001-10-31 | 2006-12-05 | Dictaphone Corporation | Distributed speech recognition system |
US7236931B2 (en) | 2002-05-01 | 2007-06-26 | Usb Ag, Stamford Branch | Systems and methods for automatic acoustic speaker adaptation in computer-assisted transcription systems |
US7292975B2 (en) | 2002-05-01 | 2007-11-06 | Nuance Communications, Inc. | Systems and methods for evaluating speaker suitability for automatic speech recognition aided transcription |
US7505901B2 (en) | 2003-08-29 | 2009-03-17 | Daimler Ag | Intelligent acoustic microphone fronted with speech recognizing feedback |
WO2010141513A3 (en) * | 2009-06-04 | 2011-03-03 | Microsoft Corporation | Recognition using re-recognition and statistical classification |
US8032372B1 (en) | 2005-09-13 | 2011-10-04 | Escription, Inc. | Dictation selection |
US9152983B2 (en) | 2005-08-19 | 2015-10-06 | Nuance Communications, Inc. | Method of compensating a provider for advertisements displayed on a mobile phone |
Families Citing this family (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR100723404B1 (ko) * | 2005-03-29 | 2007-05-30 | 삼성전자주식회사 | 음성 인식 및 반응을 위한 음성 처리 장치와 방법 |
US7822610B2 (en) * | 2005-08-09 | 2010-10-26 | Mobile Voice Control, LLC | Use of multiple speech recognition software instances |
DE102006029755A1 (de) * | 2006-06-27 | 2008-01-03 | Deutsche Telekom Ag | Verfahren und Vorrichtung zur natürlichsprachlichen Erkennung einer Sprachäußerung |
CN101853253A (zh) * | 2009-03-30 | 2010-10-06 | 三星电子株式会社 | 在移动终端中管理多媒体内容的设备和方法 |
CN107767872A (zh) * | 2017-10-13 | 2018-03-06 | 深圳市汉普电子技术开发有限公司 | 语音识别方法、终端设备及存储介质 |
CN108573707B (zh) * | 2017-12-27 | 2020-11-03 | 北京金山云网络技术有限公司 | 一种语音识别结果的处理方法、装置、设备及介质 |
KR102048030B1 (ko) | 2018-03-07 | 2019-11-22 | 구글 엘엘씨 | 자동화 어시스턴트와의 단대단 다국어 통신 촉진 |
US11354521B2 (en) | 2018-03-07 | 2022-06-07 | Google Llc | Facilitating communications with automated assistants in multiple languages |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO1998034217A1 (en) * | 1997-01-30 | 1998-08-06 | Dragon Systems, Inc. | Speech recognition using multiple recognizors |
GB2323693A (en) * | 1997-03-27 | 1998-09-30 | Forum Technology Limited | Speech to text conversion |
EP0872827A2 (de) * | 1997-04-14 | 1998-10-21 | AT&T Corp. | System und Verfahren zur distalen automatischen Spracherkennung über ein paket-orientiertes Datennetz |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0830960B2 (ja) * | 1988-12-06 | 1996-03-27 | 日本電気株式会社 | 高速音声認識装置 |
JP3265701B2 (ja) * | 1993-04-20 | 2002-03-18 | 富士通株式会社 | 多判定器によるパターン認識装置 |
JPH10177469A (ja) * | 1996-12-16 | 1998-06-30 | Casio Comput Co Ltd | 移動端末音声認識/データベース検索/リソースアクセス通信システム |
JPH10214258A (ja) * | 1997-01-28 | 1998-08-11 | Victor Co Of Japan Ltd | データ処理システム |
JP3767091B2 (ja) * | 1997-06-12 | 2006-04-19 | 富士通株式会社 | 画面対話処理装置 |
JPH1145271A (ja) * | 1997-07-28 | 1999-02-16 | Just Syst Corp | 検索条件の入力方法およびその方法の各工程をコンピュータに実行させるためのプログラムを記録したコンピュータ読み取り可能な記録媒体 |
-
1999
- 1999-03-09 DE DE19910234A patent/DE19910234A1/de not_active Withdrawn
-
2000
- 2000-02-10 KR KR1020017011408A patent/KR20010108330A/ko not_active Application Discontinuation
- 2000-02-10 CN CN00807383.XA patent/CN1350685A/zh active Pending
- 2000-02-10 WO PCT/EP2000/001145 patent/WO2000054252A2/en not_active Application Discontinuation
- 2000-02-10 AU AU26721/00A patent/AU2672100A/en not_active Abandoned
- 2000-02-10 JP JP2000604400A patent/JP2002539481A/ja active Pending
- 2000-02-10 EP EP00905058A patent/EP1163660A2/de not_active Withdrawn
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO1998034217A1 (en) * | 1997-01-30 | 1998-08-06 | Dragon Systems, Inc. | Speech recognition using multiple recognizors |
GB2323693A (en) * | 1997-03-27 | 1998-09-30 | Forum Technology Limited | Speech to text conversion |
EP0872827A2 (de) * | 1997-04-14 | 1998-10-21 | AT&T Corp. | System und Verfahren zur distalen automatischen Spracherkennung über ein paket-orientiertes Datennetz |
Cited By (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2002103675A1 (en) * | 2001-06-19 | 2002-12-27 | Intel Corporation | Client-server based distributed speech recognition system architecture |
US7133829B2 (en) | 2001-10-31 | 2006-11-07 | Dictaphone Corporation | Dynamic insertion of a speech recognition engine within a distributed speech recognition system |
US7146321B2 (en) | 2001-10-31 | 2006-12-05 | Dictaphone Corporation | Distributed speech recognition system |
WO2003049080A1 (en) * | 2001-11-30 | 2003-06-12 | Dictaphone Corporation | Distributed speech recognition system with speech recognition engines offering multiple fuctionalites |
US6785654B2 (en) | 2001-11-30 | 2004-08-31 | Dictaphone Corporation | Distributed speech recognition system with speech recognition engines offering multiple functionalities |
US7292975B2 (en) | 2002-05-01 | 2007-11-06 | Nuance Communications, Inc. | Systems and methods for evaluating speaker suitability for automatic speech recognition aided transcription |
US7236931B2 (en) | 2002-05-01 | 2007-06-26 | Usb Ag, Stamford Branch | Systems and methods for automatic acoustic speaker adaptation in computer-assisted transcription systems |
US7505901B2 (en) | 2003-08-29 | 2009-03-17 | Daimler Ag | Intelligent acoustic microphone fronted with speech recognizing feedback |
US9152983B2 (en) | 2005-08-19 | 2015-10-06 | Nuance Communications, Inc. | Method of compensating a provider for advertisements displayed on a mobile phone |
US9898761B2 (en) | 2005-08-19 | 2018-02-20 | Nuance Communications, Inc. | Method of compensating a provider for advertisements displayed on a mobile phone |
US8032372B1 (en) | 2005-09-13 | 2011-10-04 | Escription, Inc. | Dictation selection |
WO2010141513A3 (en) * | 2009-06-04 | 2011-03-03 | Microsoft Corporation | Recognition using re-recognition and statistical classification |
US8930179B2 (en) | 2009-06-04 | 2015-01-06 | Microsoft Corporation | Recognition using re-recognition and statistical classification |
Also Published As
Publication number | Publication date |
---|---|
KR20010108330A (ko) | 2001-12-07 |
CN1350685A (zh) | 2002-05-22 |
AU2672100A (en) | 2000-09-28 |
WO2000054252A3 (en) | 2000-12-28 |
JP2002539481A (ja) | 2002-11-19 |
DE19910234A1 (de) | 2000-09-21 |
EP1163660A2 (de) | 2001-12-19 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2000054252A2 (en) | Method with a plurality of speech recognizers | |
JP4597383B2 (ja) | 音声認識方法 | |
US7712020B2 (en) | Transmitting secondary portions of a webpage as a voice response signal in response to a lack of response by a user | |
JP3923513B2 (ja) | 音声認識装置および音声認識方法 | |
US6400806B1 (en) | System and method for providing and using universally accessible voice and speech data files | |
US6658414B2 (en) | Methods, systems, and computer program products for generating and providing access to end-user-definable voice portals | |
US9323848B2 (en) | Search system using search subdomain and hints to subdomains in search query statements and sponsored results on a subdomain-by-subdomain basis | |
KR101359715B1 (ko) | 모바일 음성 웹 제공 방법 및 장치 | |
US9980016B2 (en) | Video contextual advertisements using speech recognition | |
US7496497B2 (en) | Method and system for selecting web site home page by extracting site language cookie stored in an access device to identify directional information item | |
US6377927B1 (en) | Voice-optimized database system and method of using same | |
US20070106657A1 (en) | Word sense disambiguation | |
US20010054085A1 (en) | Personal voice-based information retrieval system | |
US20020052747A1 (en) | Method and system of interpreting and presenting web content using a voice browser | |
JPH113348A (ja) | 電子対話用広告装置 | |
KR20010074926A (ko) | 인터넷 브라우저 | |
WO2003039100A2 (en) | Asynchronous access to synchronous voice services | |
MX2008009455A (es) | Herramienta de busqueda que proporciona uso opcional de guias de busqueda humana. | |
US6751649B1 (en) | Server for searching for information in a network of databases | |
US20050102147A1 (en) | Method of speech-based navigation in a communications network and of implementing a speech input possibility in private information units | |
JP2003515832A (ja) | 音声ナビゲーションのためのカテゴリーによるウェブページの参照 | |
US20060075037A1 (en) | Portal for managing communications of a client over a network | |
KR20000049411A (ko) | 검색사이트의 검색결과 광고장치 및 그 방법 | |
CN107767856B (zh) | 一种语音处理方法、装置及服务器 | |
JP7512748B2 (ja) | 知識データベース生成装置及びプログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 00807383.X Country of ref document: CN |
|
AK | Designated states |
Kind code of ref document: A2 Designated state(s): AE AL AM AT AU AZ BA BB BG BR BY CA CH CN CR CU CZ DE DK DM EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG UZ VN YU ZA ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A2 Designated state(s): GH GM KE LS MW SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
AK | Designated states |
Kind code of ref document: A3 Designated state(s): AE AL AM AT AU AZ BA BB BG BR BY CA CH CN CR CU CZ DE DK DM EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG UZ VN YU ZA ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A3 Designated state(s): GH GM KE LS MW SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG |
|
DFPE | Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101) | ||
WWE | Wipo information: entry into national phase |
Ref document number: 2000905058 Country of ref document: EP |
|
ENP | Entry into the national phase |
Ref document number: 2000 604400 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 1020017011408 Country of ref document: KR |
|
WWP | Wipo information: published in national office |
Ref document number: 1020017011408 Country of ref document: KR |
|
WWP | Wipo information: published in national office |
Ref document number: 2000905058 Country of ref document: EP |
|
REG | Reference to national code |
Ref country code: DE Ref legal event code: 8642 |
|
WWW | Wipo information: withdrawn in national office |
Ref document number: 2000905058 Country of ref document: EP |
|
WWW | Wipo information: withdrawn in national office |
Ref document number: 1020017011408 Country of ref document: KR |