US20110216889A1 - Selectable State Machine User Interface System - Google Patents
Selectable State Machine User Interface System Download PDFInfo
- Publication number
- US20110216889A1 US20110216889A1 US13/044,581 US201113044581A US2011216889A1 US 20110216889 A1 US20110216889 A1 US 20110216889A1 US 201113044581 A US201113044581 A US 201113044581A US 2011216889 A1 US2011216889 A1 US 2011216889A1
- Authority
- US
- United States
- Prior art keywords
- user
- user interface
- message
- navigation
- state machine
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 230000007246 mechanism Effects 0.000 claims abstract description 55
- 230000009977 dual effect Effects 0.000 claims abstract description 7
- 238000000034 method Methods 0.000 claims description 108
- 238000004891 communication Methods 0.000 claims description 66
- 230000008569 process Effects 0.000 claims description 36
- 238000013473 artificial intelligence Methods 0.000 claims description 35
- 230000000694 effects Effects 0.000 claims description 24
- 238000012545 processing Methods 0.000 claims description 13
- 230000008859 change Effects 0.000 claims description 4
- 230000001131 transforming effect Effects 0.000 claims 3
- 230000003993 interaction Effects 0.000 claims 1
- 230000007704 transition Effects 0.000 abstract description 4
- 238000010586 diagram Methods 0.000 description 37
- 230000009471 action Effects 0.000 description 14
- 238000005516 engineering process Methods 0.000 description 14
- 230000010354 integration Effects 0.000 description 14
- 238000012552 review Methods 0.000 description 11
- 238000004458 analytical method Methods 0.000 description 10
- 238000012706 support-vector machine Methods 0.000 description 10
- 238000012549 training Methods 0.000 description 8
- 238000007726 management method Methods 0.000 description 7
- 238000003825 pressing Methods 0.000 description 7
- 238000006243 chemical reaction Methods 0.000 description 6
- 230000003287 optical effect Effects 0.000 description 6
- 230000014759 maintenance of location Effects 0.000 description 5
- 230000005540 biological transmission Effects 0.000 description 4
- 230000000881 depressing effect Effects 0.000 description 4
- 235000014510 cooky Nutrition 0.000 description 3
- 238000013461 design Methods 0.000 description 3
- 230000006855 networking Effects 0.000 description 3
- 238000009877 rendering Methods 0.000 description 3
- 238000012546 transfer Methods 0.000 description 3
- 238000013459 approach Methods 0.000 description 2
- 238000013528 artificial neural network Methods 0.000 description 2
- 230000006399 behavior Effects 0.000 description 2
- 238000013145 classification model Methods 0.000 description 2
- 230000000875 corresponding effect Effects 0.000 description 2
- 238000003066 decision tree Methods 0.000 description 2
- 230000007423 decrease Effects 0.000 description 2
- 230000001419 dependent effect Effects 0.000 description 2
- 238000001914 filtration Methods 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 230000004927 fusion Effects 0.000 description 2
- 238000010801 machine learning Methods 0.000 description 2
- 230000005055 memory storage Effects 0.000 description 2
- 230000008520 organization Effects 0.000 description 2
- 230000002093 peripheral effect Effects 0.000 description 2
- 238000012913 prioritisation Methods 0.000 description 2
- 230000000717 retained effect Effects 0.000 description 2
- 238000010845 search algorithm Methods 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 238000010200 validation analysis Methods 0.000 description 2
- 230000004075 alteration Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 238000004883 computer application Methods 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 230000002596 correlated effect Effects 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 230000003111 delayed effect Effects 0.000 description 1
- 238000012217 deletion Methods 0.000 description 1
- 230000037430 deletion Effects 0.000 description 1
- 230000000977 initiatory effect Effects 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 239000013307 optical fiber Substances 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 238000000844 transformation Methods 0.000 description 1
- 230000007723 transport mechanism Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/487—Arrangements for providing information services, e.g. recorded voice services or time announcements
- H04M3/493—Interactive information services, e.g. directory enquiries ; Arrangements therefor, e.g. interactive voice response [IVR] systems or voice portals
- H04M3/4931—Directory assistance systems
-
- A—HUMAN NECESSITIES
- A21—BAKING; EDIBLE DOUGHS
- A21B—BAKERS' OVENS; MACHINES OR EQUIPMENT FOR BAKING
- A21B5/00—Baking apparatus for special goods; Other baking apparatus
- A21B5/02—Apparatus for baking hollow articles, waffles, pastry, biscuits, or the like
- A21B5/023—Hinged moulds for baking waffles
-
- A—HUMAN NECESSITIES
- A21—BAKING; EDIBLE DOUGHS
- A21B—BAKERS' OVENS; MACHINES OR EQUIPMENT FOR BAKING
- A21B3/00—Parts or accessories of ovens
- A21B3/13—Baking-tins; Baking forms
- A21B3/133—Baking-tins; Baking forms for making bread
-
- A—HUMAN NECESSITIES
- A47—FURNITURE; DOMESTIC ARTICLES OR APPLIANCES; COFFEE MILLS; SPICE MILLS; SUCTION CLEANERS IN GENERAL
- A47J—KITCHEN EQUIPMENT; COFFEE MILLS; SPICE MILLS; APPARATUS FOR MAKING BEVERAGES
- A47J37/00—Baking; Roasting; Grilling; Frying
- A47J37/01—Vessels uniquely adapted for baking
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/50—Centralised arrangements for answering calls; Centralised arrangements for recording messages for absent or busy subscribers ; Centralised arrangements for recording messages
- H04M3/53—Centralised arrangements for recording incoming messages, i.e. mailbox systems
- H04M3/5307—Centralised arrangements for recording incoming messages, i.e. mailbox systems for recording messages comprising any combination of audio and non-audio components
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/50—Centralised arrangements for answering calls; Centralised arrangements for recording messages for absent or busy subscribers ; Centralised arrangements for recording messages
- H04M3/53—Centralised arrangements for recording incoming messages, i.e. mailbox systems
- H04M3/533—Voice mail systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M2201/00—Electronic components, circuits, software, systems or apparatus used in telephone systems
- H04M2201/39—Electronic components, circuits, software, systems or apparatus used in telephone systems using speech synthesis
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M2201/00—Electronic components, circuits, software, systems or apparatus used in telephone systems
- H04M2201/40—Electronic components, circuits, software, systems or apparatus used in telephone systems using speech recognition
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M2203/00—Aspects of automatic or semi-automatic exchanges
- H04M2203/25—Aspects of automatic or semi-automatic exchanges related to user interface aspects of the telephonic communication service
- H04M2203/251—Aspects of automatic or semi-automatic exchanges related to user interface aspects of the telephonic communication service where a voice mode or a visual mode can be used interchangeably
- H04M2203/253—Aspects of automatic or semi-automatic exchanges related to user interface aspects of the telephonic communication service where a voice mode or a visual mode can be used interchangeably where a visual mode is used instead of a voice mode
- H04M2203/254—Aspects of automatic or semi-automatic exchanges related to user interface aspects of the telephonic communication service where a voice mode or a visual mode can be used interchangeably where a visual mode is used instead of a voice mode where the visual mode comprises menus
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M2203/00—Aspects of automatic or semi-automatic exchanges
- H04M2203/45—Aspects of automatic or semi-automatic exchanges related to voicemail messaging
- H04M2203/4509—Unified messaging with single point of access to voicemail and other mail or messaging systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/42204—Arrangements at the exchange for service or number selection by voice
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/50—Centralised arrangements for answering calls; Centralised arrangements for recording messages for absent or busy subscribers ; Centralised arrangements for recording messages
- H04M3/53—Centralised arrangements for recording incoming messages, i.e. mailbox systems
- H04M3/533—Voice mail systems
- H04M3/53333—Message receiving aspects
Definitions
- This invention is related to messaging systems and more particularly to a selectable user interface system that can employ variable interactive mechanisms to navigate through and control an underlying application state machine.
- a state machine is a computing device designed with the operational states required to perform a particular task or to solve a specific problem. There are countless special-purpose devices built as state machines. A voicemail system is an example of such an application state machine.
- DTMF dual tone multi-frequency
- DTMF systems employ a navigational technique that allows a user to spell words via a telephone keypad. This is often cumbersome and very time consuming.
- Other variations of DTMF voicemail systems employ voice media processing components that present to a user a set of available options. For example, a user is frequently addressed with a specific prompt such as, “please enter your password” or “press one to review all new messages.” While these audible prompts are sometimes useful, they tend to be time consuming and, nonetheless, often still require the user to employ DTMF techniques via a keypad.
- Recent efforts have been directed to utilizing speech recognition techniques to enable a user to navigate through state machine menus. These efforts have been very limited in that they only recognize preprogrammed menu selections and/or data. For example, a system can be programmed to recognize the names of individuals employed by a company.
- state machines today are limited to employing a single navigational technique at any given time.
- a user initially chooses to navigate through a state machine via DTMF or voice command techniques.
- both options are usually not simultaneously or concurrently available and the user does not have the option of toggling between the navigational systems during any one session.
- the subject invention in one aspect thereof, allows a user to connect to a server (e.g., messaging server) via a telephonic communication component thereby accessing the rich information therein.
- the invention comprises a system and/or methodology that unifies a telephonic communication component and/or system with an application state machine (e.g., messaging server component and/or system). More particularly, in one aspect, the invention creates two relatively distinct user experiences, one through the use of dual tone multi-frequency (DTMF) navigation and one through speech recognition navigation. In accordance thereto, one single underlying state machine can be used.
- DTMF dual tone multi-frequency
- Navigation and flow control e.g., state transitions
- UI user interface
- the invention introduces speech recognition features together with other input mechanisms to drive the user interface of an application state machine (e.g., a unified messaging system (UM)).
- UM unified messaging system
- the speech recognition UI is designed to provide a natural navigation through the application independent of a DTMF user interface. The user can navigate through the speech menu without having to enter any DTMF data.
- both the DTMF-based menu infrastructure and the speech recognition menu structure can be active and available to the user at any time. For example, at any time during the experience the user can switch from speech mode to DTMF driven mode. Because the switch does not need to be explicit, the user can issue either a DTMF or a speech command at any time.
- the prompts that will be played for the user upon entering any part of the application can be based on the UI mechanism that the user used to enter the state.
- the speech menu will be played if speech commands brought the user to the state.
- the DTMF menu will be played if a DTMF command brought the user to the state.
- the user can, at any time, say “help” to review available speech commands or press “0” to review available DTMF commands.
- both sets of commands are available at all times.
- the default UI that is presented to the user on first entry to the UM system can be set as a personal option.
- Rule-based logic and/or artificial intelligence reasoning mechanisms can be employed to determine the default UI.
- the underlying application state machine is leveraged for the multiple co-existent UI models.
- Specific to aspects of the speech UI are additional design paradigms selected to exploit some unique features of speech recognition.
- the speech UI can have distinct prompts and menus for handling user input, input validation, and error handling for low confidence recognition results.
- a set of “anytime” shortcuts can be available at all times to jump to the main functionality of the underlying state machine (e.g., unified messaging application).
- these main functionality locations can be voicemail, calendar, call someone, email, and personal options.
- Each speech menu can also have menu commands that are specific to the menu context.
- the invention can add functionality to enable telephonic generation and retrieval of email communications.
- the invention can further facilitate accepting or canceling calendar items (e.g., meeting requests and appointments) via a telephonic communications device.
- FIG. 1 is a high-level exemplary system architecture that facilitates state machine user interface selection in accordance with an aspect of the invention.
- FIG. 2 is an exemplary system that illustrates a user interface store having disparate state machine user interface components in accordance with a disclosed aspect.
- FIG. 3 illustrates a system architecture that facilitates user interface selection with respect to a unified messaging system in accordance with a disclosed aspect.
- FIG. 4 illustrates a user interface selection component that employs a rule-based logic engine in accordance with a disclosed aspect.
- FIG. 5 illustrates a user interface selection component that employs an artificial intelligence component in accordance with an exemplary aspect.
- FIG. 6 illustrates a high-level exemplary system architecture that facilitates communication in accordance with an aspect of the subject invention.
- FIG. 7 illustrates an exemplary system architecture that employs a unified messaging system in accordance with a disclosed aspect.
- FIG. 8 is a call control component that facilitates content conversion and interpretation in accordance with a disclosed aspect.
- FIG. 9 illustrates a flow diagram of an exemplary communication methodology in accordance with a disclosed aspect.
- FIG. 10 illustrates a call control component that employs a rule-based logic engine in accordance with a disclosed aspect.
- FIG. 11 illustrates a call control component that employs an artificial intelligence component in accordance with an aspect.
- FIG. 12 is an exemplary flow diagram of calling into a unified messaging system in accordance with a disclosed aspect.
- FIG. 13 is an exemplary flow diagram of accessing voicemail and email in accordance with an aspect.
- FIG. 14 is an exemplary flow diagram of accessing an electronic calendar in accordance with an exemplary aspect.
- FIG. 15 is an exemplary flow diagram of accessing electronic contacts in accordance with a disclosed aspect.
- FIG. 16 is an exemplary flow diagram of recording a message in accordance with a disclosed aspect.
- FIG. 17 is an exemplary flow diagram of calling into a unified messaging system that employs voice commands in accordance with a disclosed aspect.
- FIG. 18 is an exemplary flow diagram of accessing voicemail and email that employs voice commands in accordance with an aspect of the invention.
- FIG. 19 is an exemplary flow diagram of accessing an electronic calendar that employs voice commands in accordance with a disclosed aspect.
- FIG. 20 is an exemplary flow diagram of accessing electronic contacts that employs voice commands in accordance with a disclosed aspect.
- FIG. 21 is an exemplary flow diagram of recording a message that employs voice commands in accordance with a disclosed aspect.
- FIG. 22 illustrates a block diagram of a computer that can execute the disclosed architecture.
- FIG. 23 illustrates a schematic block diagram of an exemplary computing environment in accordance with the subject invention.
- a component can be, but is not limited to being, a process running on a processor, a processor, an object, an executable, a thread of execution, a program, and/or a computer.
- an application running on a server and the server can be a component.
- One or more components can reside within a process and/or thread of execution, and a component can be localized on one computer and/or distributed between two or more computers.
- the term to “infer” or “inference” refer generally to the process of reasoning about or inferring states of the system, environment, and/or user from a set of observations as captured via events and/or data. Inference can be employed to identify a specific context or action, or can generate a probability distribution over states, for example. The inference can be probabilistic—that is, the computation of a probability distribution over states of interest based on a consideration of data and events. Inference can also refer to techniques employed for composing higher-level events from a set of events and/or data. Such inference results in the construction of new events or actions from a set of observed events and/or stored event data, whether or not the events are correlated in close temporal proximity, and whether the events and data come from one or several event and data sources.
- the system 100 can include a user interface (UI) selection component 102 , a UI store 104 and a state machine component 106 .
- UI user interface
- a novel aspect of the invention is to decouple the UI mechanisms from the state machine 106 .
- the invention can employ multiple UI mechanisms to effect navigation within an application state machine.
- a novel aspect of the invention can concurrently employ multiple UI interfaces (e.g., dual tone multi-frequency (DTMF), speech recognition).
- DTMF dual tone multi-frequency
- multiple UI mechanisms can be active at any time allowing a user to freely switch between UI mechanisms.
- this invention employs a novel design that allows for the co-existence of two or more UI mechanisms (e.g., DTMF and speech recognition) that enable telephone access to a user inbox.
- the two or more UI mechanisms can simultaneously leverage the underlying application state machine.
- the exemplary aspects that follow are directed to the use of the UI selection component 102 and the UI store 104 in connection with a voicemail application (e.g., state machine 106 ).
- a voicemail application e.g., state machine 106
- the state machine component 106 can be any application state machine.
- the UI selection component 102 can effect utilization of any available UI mechanisms.
- a user can employ the UI selection component 102 to select a DTMF navigational technique to control the operation of an underlying state machine (e.g., voicemail system).
- the UI selection component 102 can be employed to facilitate a speech recognition navigational technique.
- the UI selection component 102 can be continuously available thereby providing a user the ability to toggle between UI navigational techniques at any time during a session.
- FIG. 2 there is illustrated a more detailed architectural diagram of an aspect of the invention.
- state machine component 106 can be used in connection with multiple UI mechanisms.
- flow control e.g., state transitions
- UI mechanisms e.g., state transitions
- the UI store 104 can include 1 to N UI components, where N is an integer. It will be appreciated that 1 to N UI components can be referred to individually or collectively as UI components 202 . A more specific example of the UI store 104 is illustrated in FIG. 3 .
- the UI store 104 can include UI components 202 including, but not limited to, a DTMF UI component and a speech UI component.
- the state machine component 106 can include a unified messaging system 302 .
- the unified messaging system 302 can unite a telephonic system with an electronic messaging and/or organization system. An exemplary unified messaging system will be discussed in detail infra.
- the invention introduces speech recognition features together with DTMF features to drive the UI of unified messaging system 302 .
- the speech recognition UI 202 can be designed to provide a natural navigation through the application (e.g., state machine 106 ) independent of the DTMF UI 202 . Because the invention decouples these UI mechanisms, the user will be able to drive through the speech menu without having to enter any DTMF input(s).
- the user can independently switch from speech mode to DTMF driven mode.
- the switch need not be explicit; the user need only issue either a DTMF signal or a speech command to automatically effect the transition.
- the prompts that will be played for the user on entering any part of the application will be based on the UI mechanism the user employed to enter the state. If speech commands were employed to bring the user to a state, the speech menu will be played. On the other hand, if a DTMF command brought the user to the state, the DTMF menu will be played. Additionally, from either the speech or DTMF UI 202 the user can employ “help” to determine the commands. For example, at any time the user can say “help” to review available speech commands or press “0” to review available DTMF commands. Both sets of commands are available at all times.
- a default UI that is presented to the user on first entry to the unified messaging system 302 can be set as a personal option.
- rule-based logic and/or artificial intelligence (AI) techniques can be employed to automate the selection process. These rule-based logic and AI based reasoning techniques will be discussed in greater detail infra.
- the speech UI 202 can have distinct prompts and menus for handling user input, input validation, and error handling for low confidence recognition results. Further, the DTMF prompts will not be reused or converted to “press or say”. The speech prompts and menus can go through a full usability pass including user education review. Menus can be generated to sound natural for the user. For example, a menu can be generated to say “Would you like to make a call or access your mailbox?” rather than “Say ‘make a call’ to call someone or ‘access mailbox’ to access your mailbox”.
- a set of “anytime” shortcuts can be available at all times to jump to the main functionality of the unified messaging application 302 .
- the main functionality can include voicemail, calendar, call someone, email, and personal options.
- Each speech menu can also have menu commands that are specific to the menu context.
- the invention can limit access to features of the unified messaging application 106 .
- some more obscure or less important features may only be provided through DTMF mode.
- These features can be preprogrammed and/or defined by a user as desired.
- menu flexibility can be increased since there is no concern around reusing DTMF digits. For example, in the voicemail, email and calendar playback menus there does not need to be a distinction between the “during playback” and the “after message” menu.
- FIGS. 17 to 21 display both the DTMF menu infrastructure and the speech user interface overlay.
- the speech diagrams are intended to show the main grammar elements of an exemplary UI and the overall flow rather than every specific prompt and error handling condition.
- the flow diagrams included are not intended to limit the scope and/or novel functionality of the invention in any way.
- novel aspects of the invention can employ rule-based logic and/or AI reasoning technology to effect, predict and/or infer an action. More particularly, it will be appreciated that the subject invention (e.g., in connection with state analysis, UI selection, voice recognition) can employ various rule-based and/or AI based schemes for carrying out various aspects thereof.
- UI selection component 102 can include a rule-based logic engine 402 .
- an optional AI component (not shown) can be used together with, or in place of, the rule-based logic engine 402 to automatically infer an action or set of actions to be employed in connection with the functionality of the UI selection component 102 described supra.
- the rule-based logic engine 402 can be programmed or configured in accordance with a predefined preference (e.g., a rule).
- a rule e.g., logic
- the system can automatically select and employ a specific UI.
- a rule can be established to take into consideration the origination location (e.g., internal, external, Internet . . . ), originator, target location, etc., to automatically select and employ a UI.
- FIG. 5 A schematic diagram of another alternative aspect of the subject invention is illustrated in FIG. 5 .
- the UI selection component 102 of this alternative aspect illustrated in FIG. 5 employs an AI component 502 that can automatically infer and/or predict an action.
- This alternative aspect can optionally include an inference module (not shown) that facilitates automatic selection of the UI selection component 102 .
- the optional AI component 502 can facilitate automatically performing various aspects (e.g., state analysis, UI selection, and voice recognition) of the subject invention as described herein.
- the AI component 502 can optionally include an inference component (not shown) that can further enhance automated aspects of the AI component utilizing, in part, inference-based schemes to facilitate inferring intended actions to be performed at a given time and/or state.
- the AI-based aspects of the invention can be effected via any suitable machine-learning based technique and/or statistical-based techniques and/or probabilistic-based techniques.
- the subject invention can optionally employ various AI-based schemes for automatically carrying out various aspects thereof.
- the AI component 502 can optionally be provided to implement aspects of the subject invention based upon AI processes (e.g., confidence, inference . . . ). For example, a process for initiating a UI based upon user preferences or sender identification can be facilitated via an automatic classifier system and process. Further, the optional AI component 502 can be employed to facilitate an automated process of selecting a UI based upon a predicted and/or inferred preference.
- Such classification can employ a probabilistic and/or statistical-based analysis (e.g., factoring into the analysis utilities and costs) to prognose or infer an action that a user desires to be automatically performed.
- a support vector machine is an example of a classifier that can be employed.
- the SVM operates by finding a hypersurface in the space of possible inputs, which hypersurface attempts to split the triggering criteria from the non-triggering events. Intuitively, this makes the classification correct for testing data that is near, but not identical to training data.
- Other directed and undirected model classification approaches include, e.g., naive Bayes, Bayesian networks, decision trees, and probabilistic classification models providing different patterns of independence can be employed. Classification as used herein also is inclusive of statistical regression that is utilized to develop models of priority.
- the subject invention can employ classifiers that are explicitly trained (e.g., via a generic training data) as well as implicitly trained (e.g., via observing user behavior, receiving extrinsic information).
- SVM's can be configured via a learning or training phase within a classifier constructor and feature selection module.
- expert systems, fuzzy logic, support vector machines, greedy search algorithms, rule-based systems, Bayesian models (e.g., Bayesian networks), neural networks, other non-linear training techniques, data fusion, utility-based analytical systems, systems employing Bayesian models, etc. are contemplated and are intended to fall within the scope of the hereto appended claims.
- the system 600 can include a unified messaging (UM) system component 602 , a telephonic communication component 604 and a computer application, state machine and/or server component 606 .
- UM unified messaging
- the UI selection component 102 FIG. 1
- the subject invention can be employed to enable a UI to be selected thus effecting telephonic access to rich data maintained within a server component. This telephonic access can be effected via a unified messaging system 602 .
- the telephonic communication component 604 can include any device capable of communicating voice and/or audible sounds and signals (e.g., DTMF).
- the telephonic communication component 604 can include a traditional wired telephone as well as wireless (e.g., cellular) telephones.
- the telephonic communication component 604 can be effected via a portable wireless device and/or computer that employs voice over Internet protocol (VOIP) or fax over Internet protocol (FOIP).
- VOIP voice over Internet protocol
- FOIP fax over Internet protocol
- the server component 606 can be employed to host any communication and/or organizational application.
- the server component 606 can host communication applications including, but not limited to, voicemail, email, text messaging applications or the like in connection with aspects of the subject invention.
- the UM system 602 can unite the telephonic communications system 604 with the server component 606 (e.g., mailbox server). Accordingly, the telephonic communications system 604 can access the rich information maintained in the server component 606 thereby, unifying the systems.
- FIG. 6 illustrates a single telephonic communications system 604 and a single server component 606 , it is to be understood and appreciated that the invention is extensible thereby being capable of uniting multiple telephonic and/or server systems.
- the system 700 generally includes the UM system 602 , telephonic communications system 604 and server component 606 .
- the telephonic communication component 604 , the server component 606 and a plurality of clients 702 can be connected to the UM system 602 via a wired local area network (LAN) connection 704 (e.g., Ethernet).
- LAN local area network
- FIG. 7 connections illustrated in FIG. 7 can be wired, wireless or a combination thereof.
- the clients 702 can employ disparate communication techniques in order to communicate with the server component 606 .
- a smart phone, pocket personal computer (PC) or other handheld device can employ specialized communication software to synchronize to the mailbox server 706 or the active directory server 708 .
- a desktop e.g., client 702
- a wired connection e.g., Ethernet
- Yet another exemplary client of the three clients 702 e.g., laptop
- a call control component 710 text-to-speech (TTS) engine 712 , speech recognition engine 714 , voice media processing component 716 , fax media processing component 718 and UM application component 720 can be provided.
- TTS text-to-speech
- the telephonic communications component 604 can include a public telephone 722 connected via a public switched telephone network (PSTN) 724 , a private (e.g., in-house company) phone 726 connected via a private branch exchange (PBX) 728 or private telephone switchboard, and computer system 730 that employs VOIP or FOIP via gateway 732 . It will be appreciated that these components are exemplary and are not intended to be an exhaustive list of the telephonic communications component 604 . In other words, telephonic communications component 604 can include any mechanism capable of generating and/or transmitting audible (e.g., voice, fax) signals.
- audible e.g., voice, fax
- UM system 602 can include a UI selection component 734 to effect selection of an appropriate UI as described supra.
- the UI selection component 734 can communicate with a UI store (not shown) in order to effect the desired and/or selected UI.
- the UI store can be co-located or located remotely from the UI selection component 734 .
- the UI store (not shown) can be located within server component 606 .
- FIG. 8 illustrates an implementation of the call control component 710 .
- UM application component 720 can include an email integration component 802 , a calendar integration component 804 , a contact list integration component 806 and a security component 808 .
- Each of these components 802 , 804 , 806 , 808 ) will be better understood upon a discussion of the exemplary scenarios that follow.
- the UM application component 720 of FIG. 8 illustrates four exemplary functional components, it is to be appreciated that the functionality of these components described in the following scenarios can be combined (or separated) into disparate components.
- the components described herein can be co-located or remotely located without departing from the scope of the subject invention.
- a first exemplary scenario is directed to a discussion of the email integration component 802 . More particularly, this scenario is directed to integration of an “out-of-office” (OOF) voicemail greeting and email.
- OEF out-of-office
- the OOF status can be separately set on both a user specific email account (e.g., via client 702 ) or voicemail account (e.g., via telephone 722 , 726 ).
- an OOF email auto-reply generated via the server component as well as the voicemail extended absence (e.g., OOF) greeting often contain similar information.
- the subject invention can facilitate setting both the email and the voicemail OOF status from a single location in a single operation.
- the UM system 602 can integrate (e.g., synchronize) the OOF message into the telephonic communications component 604 . Accordingly, calls to the user via the telephonic communications system 604 will include an audible version of the OOF extended absence greeting. Through the use of text-to-speech conversion techniques of the TTS engine 712 , this greeting can audibly render the text contained in the email OOF auto-reply.
- the unified messaging system 602 can automatically set an email OOF message in the mailbox server 706 . Thereafter, a sender of an email to the user will receive a text version of the OOF auto-reply that was originated via telephonic communication mechanisms.
- a second scenario is directed to the calendar integration component 804 and particularly to calendar or planner access and management.
- users have limited access to their calendar from telephonic systems (e.g., 722 , 726 , 730 ) and can only read the information corresponding to appointments, which is maintained within server component 606 .
- users can manage appointments and initiate communications related to appointments via the telephonic communications component 604 (e.g., telephone 722 , 726 , 730 ).
- full planner functionality can be facilitated via speech recognition component 714 of the UM system 602 .
- a user can accept, decline, modify, cancel, and communicate with respect to appointments contained within the server component 606 .
- a telephonic session with the UM system 602 can be automatically transferred to a telephone of a meeting location.
- a user via the telephonic communications component 604 , a user can connect to the UM system 602 whereby, based upon a calendar (e.g., appointment) entry in the server component 606 , a determination can be made with respect to an appropriate transfer target location for the call.
- a user can initiate a targeted or broadcast electronic message (e.g., email) to the meeting participants. For example, a user can inform meeting participants of a delayed arrival to a meeting.
- a targeted or broadcast electronic message e.g., email
- Yet another aspect is directed to remote access of contact list entries via the contact list integration component 806 .
- users cannot telephonically access their personal contact list, which is frequently stored on the email system (e.g., server 606 ).
- the email system e.g., server 606
- a user can search for contacts in the contact list maintained within the server 606 . If desired, the user can immediately connect to or leave a message (e.g., voice, text, email) for the contact. If the contact is an external contact with only an email address, this invention can allow a user to communicate an audible message by creating an email with a sound file attachment (e.g., .wav) thereby forwarding the communication to the intended target.
- a sound file attachment e.g., .wav
- the subject invention can be employed to reset a personal identification number (PIN) via an email or other application.
- PIN personal identification number
- this invention can allow a user to log into their email on the server component 606 using domain credentials and reset/change their telephone voicemail PIN (e.g., password).
- Yet another scenario involving the security component 808 is directed to digital rights management for voicemail.
- voice messages can be sent with restricted access. For instance, a voicemail can be sent whereby a recipient can listen to the voicemail, but can be restricted from forwarding or saving the message.
- the UM system 602 (e.g., email integration component 802 ) of the subject invention provides for text preview of voicemails (and vice versa).
- users via an email client 702 , users can textually access voicemails on their computer, laptop or other suitable device (e.g., smart phone).
- voice recognition techniques e.g., speech recognition engine 714
- the subject invention can convert the voicemail thereby transcribing it into text. Therefore, users can review the content of the voice message without playing the sound file.
- the TTS engine 712 can be employed to convert and deliver email communications via the telephonic communications device 604 .
- the invention can facilitate converting electronic message attachments (e.g., word processing documents) to audible files (e.g., voice). This can be accomplished via the TTS component 712 .
- the subject invention in addition to rendering the audible content of an electronic communication, the subject invention can render an audible file that contains the content of a textual attachment.
- the speech recognition engine 714 can be employed to effect navigation within a file system directory.
- the TTS engine 712 can be employed to convert a target document to speech thus rendering the audible (e.g., voice) file to a user via the telephonic communications component 604 .
- Another aspect is directed to employing the email integration component 802 to facilitate analyzing the content of an email communications message. For example, when email messages are audibly rendered to a user via the telephonic communications device 604 , (e.g., telephone), often the messages can be very long and difficult to follow.
- a novel feature of an aspect of this invention can employ logic and/or reasoning techniques to identify the most important messages and/or sentences (e.g., content) included in a communication. These logic and/or reasoning mechanisms can employ rule-based and/or artificial intelligence (AI) mechanisms to effect filtering and/or sorting of the content. Once identified, a streamlined or condensed version of the content can be audibly rendered to a user.
- AI artificial intelligence
- the UM system 602 can limit the audible playback of email messages to include only the most important messages and/or sentences included therein. It will be appreciated that because the content can be filtered and/or sorted in accordance with a desired algorithm (e.g., rule based, AI), the quantity and length of the message(s) that are read to the user can be decreased. Likewise, important voicemail messages can be handled (e.g., filtered, sorted) in the same manner. Of course, when people receive many email messages over the telephone it can be difficult to navigate to the particular messages that are important to the user. With this invention, important voicemails can be identified by the UM system 702 and read to the user first. As well, voice messages can be filtered whereby only portions of the voicemails are read to the user.
- a desired algorithm e.g., rule based, AI
- Another novel feature of the subject invention is directed to facsimile transmissions.
- all inbound facsimiles are directed to one mailbox where a router looks at the facsimile and directs it, via electronic means, to the appropriate person in the organization. If the facsimile message contains sensitive information, the person who is redirecting the facsimile would have access to information that was intended to be confidential.
- the UM system 602 can employ the fax media processing component 718 to separate the cover page from the remaining pages of a facsimile transmission. Therefore, the router who is redirecting the message can only have access to the cover page of the facsimile and not the confidential contents.
- the UM system 602 can generate and forward an email message to the callee thereby notifying that someone has called but did not leave a voicemail message.
- Still another novel aspect of the invention is directed to policy and compliance integration with an email system.
- voice messages are stored on a server (e.g., server component 606 ) and are retained and/or archived according to an arbitrary or preprogrammed policy.
- a server e.g., server component 606
- one such policy can delete messages after a specific number of days (e.g., 21).
- a user can control which messages are deleted or archived from the telephonic communications component 604 .
- logic and/or reasoning mechanisms can be applied to automatically manage retention and/or deletion of messages.
- Yet another aspect of the invention is directed to a system and/or methodology to permit a user to select a keystroke profile scheme.
- voicemail systems can have different keys to navigate and control the telephone user interface. If users change from one system to another, they often have to learn the new keys to use the new system.
- button profiles can be employed for the most commonly used systems. Accordingly, a user can choose a predetermined profile that is familiar to that particular user. As well, buttons can be programmed as desired by a specific user. In this manner, the user does not necessarily have to learn a new keystroke profile scheme to access a new system.
- Another novel aspect of the security component 808 is directed to employing voice recognition techniques to effect voicemail authentication.
- voice recognition techniques Today, when users access their mailbox from the telephone, they typically key in a PIN as their password. In many cases, this is not very secure and is inconvenient to the user (e.g., accessing while driving a car).
- users can audibly speak a phrase or desired PIN to gain access to their mailbox. This can be more convenient and also more secure as the voice phrase must match a previously recorded phrase by the user.
- the system can be programmed to learn qualities (e.g., voice tonal qualities) specific to a user.
- Yet another exemplary aspect is directed to end-user configuration of menus and prompts.
- end-users access their mailbox via the telephone, they do not have the capability to change the buttons or the information that is played back to them during the greeting.
- users can specify a location of their greeting, therefore, depending on the user role, a more detailed greeting that contains additional information may be used. For example, an IT administrator might choose to put the number of critical escalation cases in their main mailbox greeting.
- prompts e.g., audio
- the ordering of the state machine can user customizable.
- aspects of the subject invention are directed toward at least the following novel components employed in connection with the UM system 602 .
- OOF Voicemail Greeting Integration with Email Application The aspect that the email and voicemail systems OOF status can be content matched thereby enabling a user to set the OOF status from one place.
- the auto-reply email body can be read to the user as the telephone greeting using TTS mechanisms.
- Calendar Access Capability to act on and interact with calendar appointments from the telephone in a novel manner. For example, a user can accept, decline, cancel, transfer to the phone of the meeting location, and initiate an email message to the meeting participants.
- Contact List Access Capability to access personal contact list, rather than just the company directory. Ability to send, via the telephone, an email and sound file attachment to an email account.
- PIN Reset Allowing the user to reset a password (e.g., voicemail password) via an email client. It will be understood that the operating system security credentials can maintain security.
- Digital Rights Management for Voicemail are a new and novel concept.
- the invention can be employed to control and/or restrict access to voicemail content. For instance, a recipient can listen to a voicemail, but can be restricted from forwarding or saving the message.
- Text Preview of Voicemails Using speech to text conversion techniques, the subject invention can convert recorded voicemails into a textual message. The message can then be sent to the recipient via an email client application. As well, the system contemplates text-to-speech conversion for text messages. Moreover, the system contemplates generating and attaching a sound file (e.g., .wav format) of a voice message to an email. Thus, the email and attachment can be forwarded to the recipient.
- a sound file e.g., .wav format
- the subject invention can extract the cover page (e.g., first page) from a received facsimile transmission thereby, retaining confidentiality of remaining pages of the transmission.
- cover page e.g., first page
- the invention can determine and rank sentences of a communication (e.g., email and/or voicemail) in terms of importance. For example, ranking can occur with regard to a given an email thread and key words contained therein. It will be appreciated that the ranking of sentences and/or keywords can generally provide a summary of the entire email without having to review it completely. Additionally, aspects are directed to using this technology for reading textual emails via the telephone, which is also a novel concept of the invention.
- a communication e.g., email and/or voicemail
- filtering criteria e.g., rule-based, AI
- the subject invention can analyze email threads and rank them accordingly.
- the system can determine the sender of a message, employ a policy, logic and/or reasoning mechanism thereby ranking and/or categorizing the importance of the messages.
- the system can use this technology to determine an order with respect to reading the messages to the user over the telephone.
- this technology can be employed to filter emails prior to presenting them to a user via a telephone.
- Missed Call Notification This component can be integrated into the unified messaging system thereby providing third party call control by monitoring traffic.
- the system can monitor and advise a client of received calls.
- Button Profiles The subject invention provides for functionality that integrates novel button profile definition mechanisms.
- the invention can employ a secure spoken password to access a voicemail and/or email mailbox through a telephone. This audible authentication is a novel use of the authentication technology.
- the subject invention provides for the ability to customize end-user menus and prompts.
- FIG. 9 illustrates an exemplary flow chart of acts to employ an OOF voice message to set an email OOF notification. While, for purposes of simplicity of explanation, the one or more methodologies shown herein, e.g., in the form of a flow chart or diagram, are shown and described as a series of acts, it is to be understood and appreciated that the subject invention is not limited by the order of acts, as some acts may, in accordance with the subject invention, occur in a different order and/or concurrently with other acts from that shown and described herein. For example, those skilled in the art will understand and appreciate that a methodology could alternatively be represented as a series of interrelated states or events, such as in a state diagram. Moreover, not all illustrated acts may be required to implement a methodology in accordance with the subject invention.
- an audible OOF message is recorded.
- the system can identify corresponding email account(s) at 904 . Once identified, conversion can begin.
- the audible OOF message can be converted to a text OOF message.
- the text OOF message can be applied to the identified email account(s).
- the methodology of FIG. 9 illustrates a process that can convert a voicemail OOF into an email OOF, it is to be understood that an alternate novel aspect of the subject invention can be employed to convert an email OOF into a voicemail system.
- novel aspects of the invention can employ rule-based logic and/or AI reasoning technology to effect, predict and/or infer an action. More particularly, it will be appreciated that the subject invention (e.g., in connection with content analysis, content retention policy, synchronization, voice recognition) can employ various rule-based and/or AI-based schemes for carrying out various aspects thereof.
- UM application component 720 can include a rule-based logic engine 1002 .
- an optional AI component (not shown) can be used together with, or in place of, the rule-based logic engine 1002 to automatically infer an action or set of actions to be employed in connection with the functionality of the UM application component 720 (and UM system 602 ) described supra.
- the rule-based logic engine 1002 can be programmed or configured in accordance with a predefined preference (e.g., rule).
- a rule e.g., logic
- the emails can be converted and read to a user in a prioritized order.
- a rule can be established to take into consideration the subject, sender identification, recipient identification, etc., to prioritize and process emails.
- a rule can be established to predetermine or define a policy whereby emails and/or voicemails can be managed (e.g., retained, forwarded, deleted) in accordance with the policy.
- FIG. 11 A schematic diagram of another alternative aspect of the subject invention is illustrated in FIG. 11 .
- the UM application component 720 of this alternative aspect illustrated in FIG. 11 employs an optional AI component 1102 that can automatically infer and/or predict an action.
- This alternative aspect can optionally include an inference module (not shown) that facilitates automatic control and/or operation of the UM application component 720 .
- the optional AI component 1102 can facilitate automatically performing various aspects (e.g., analysis and prioritization of content, content retention policy, synchronization, voice recognition) of the subject invention as described herein.
- the AI component 1102 can optionally include an inference component (not shown) that can further enhance automated aspects of the AI component utilizing, in part, inference based schemes to facilitate inferring intended actions to be performed at a given time and/or state.
- the AI-based aspects of the invention can be effected via any suitable machine-learning based technique and/or statistical-based techniques and/or probabilistic-based techniques.
- the subject invention can optionally employ various artificial intelligence based schemes for automatically carrying out various aspects thereof.
- the AI component 1102 can optionally be provided to implement aspects of the subject invention based upon AI processes (e.g., confidence, inference). For example, a process for determining the prioritization of content in an email based upon user preferences or sender identification can be facilitated via an automatic classifier system and process. Further, the optional AI component 1102 can be employed to facilitate an automated process of rendering prioritized content to a user based upon a predicted and/or inferred preference.
- Such classification can employ a probabilistic and/or statistical-based analysis (e.g., factoring into the analysis utilities and costs) to prognose or infer an action that a user desires to be automatically performed.
- a support vector machine is an example of a classifier that can be employed.
- the SVM operates by finding a hypersurface in the space of possible inputs, which hypersurface attempts to split the triggering criteria from the non-triggering events. Intuitively, this makes the classification correct for testing data that is near, but not identical to training data.
- Other directed and undirected model classification approaches include, e.g., naive Bayes, Bayesian networks, decision trees, and probabilistic classification models providing different patterns of independence can be employed. Classification as used herein also is inclusive of statistical regression that is utilized to develop models of priority.
- the subject invention can employ classifiers that are explicitly trained (e.g., via a generic training data) as well as implicitly trained (e.g., via observing user behavior, receiving extrinsic information).
- SVM's can be configured via a learning or training phase within a classifier constructor and feature selection module.
- expert systems fuzzy logic, support vector machines, greedy search algorithms, rule-based systems, Bayesian models (e.g., Bayesian networks), neural networks, other non-linear training techniques, data fusion, utility-based analytical systems, systems employing Bayesian models, etc., are contemplated and are intended to fall within the scope of the hereto appended claims.
- FIGS. 12 to 16 Illustrated in FIGS. 12 to 16 are flow diagrams of an exemplary user interface (UI) of a UM system in accordance with an aspect of the invention.
- UI user interface
- this UM system can facilitate management of voicemail, email and faxes stored in a server mailbox.
- the subject invention facilitates access to and management of this information from a telephone UI.
- users are able to access their voicemail and other useful information from the telephone in lieu of accessing a desktop computer or handheld device.
- FIGS. 12 to 16 are illustrative examples of a DTMF process flow in accordance with a disclosed aspect.
- FIGS. 17 to 21 are illustrative examples of a speech activated process flow in accordance with a disclosed aspect.
- the DTMF and speech recognition interfaces can be active concurrently with respect to UM system navigational mechanisms.
- this invention can address.
- a user is headed to a meeting in a remote location, but is not sure which room in the building the meeting will be held. The user may want to quickly find this information while driving his car toward the meeting.
- One aspect of this invention can facilitate telephonic retrieval of this information.
- a third scenario suppose a user is driving to work in the morning and wants to get a head start on the some of the issues that are pending in emails or voicemails. From the telephone UI, the user can quickly review the different issues, reply as desired, and follow-up with the people involved with the issues.
- FIG. 12 an exemplary process task flow diagram of calling into a UM is shown.
- the process illustrates a flow of procedures in accordance to a call initiated by an external caller 1202 and an internal caller 1204 .
- the process flow of the exemplary aspect is dependent upon the type of caller (e.g., external 1202 or internal 1204 ) as well as the contact number dialed at 1206 .
- An external caller will be prompted with the company information at 1208 whereby an internal caller with internal information at 1210 .
- a caller can contact a user located on the system by spelling the name and employing the directory search at 1212 or by connecting directly to the extension ( 1214 ) by entering the numeric extension. If the user does not answer, at 1216 , the system can play the appropriate greeting (e.g., external, internal, OOF) in accordance with the type of caller. Accordingly, a voicemail message can be recorded at 1218 . If a pre-designated key (e.g., “*”) is pressed, the system can enter the callee mailbox at 1220 .
- the appropriate greeting e.g., external, internal, OOF
- a voicemail message can be recorded at 1218 .
- a pre-designated key e.g., “*”
- the system can enter the callee mailbox at 1220 .
- the callee can access the mailbox directly from 1220 by choosing a pre-designated key (e.g., “#”) from the company and/or internal greeting menu ( 1208 , 1210 ). This direct access will prompt the callee for necessary information at 1222 . In each case, once the appropriate security credential information is entered at 1220 , access to the UM is gained at 1224 .
- FIG. 13 illustrates a continued procedure flow diagram once access is granted.
- the user can be prompted with a main menu at 1302 .
- the user can be prompted to press “1” for voicemail, “2” for calendar options, “3” for contacts, “4” for email and so on.
- Option numbers “2” and “3”, e.g., calendar options and contacts, will be discussed in greater detail with reference to FIGS. 14 and 15 respectfully.
- a user can choose “5” to send a message (e.g., voicemail, email) or “7” to set an OOF. Additional exemplary menu choices are illustrated in block 1302 .
- voicemail or email the system proceeds to 1304 whereby content can be presented to the user.
- the user can be presented with voicemail, email and/or meeting request content.
- logic and reasoning mechanisms can be employed to manage, organize, filter and/or sort the content as described supra.
- the user can be presented with an “after message” menu thereby presenting additional options.
- by pressing “4” a user can send a facsimile of a desired message.
- the user can opt to send a facsimile of the message to a desired location.
- the system can play the number of pending meetings.
- the user can manage the meeting appointments and/or requests by selecting any of various options. For example, the user can select option “2” to call the meeting location, “2,2” to call the meeting organizer or “3” to indicate a late arrival to the subject meeting. As well, additional navigational options can be presented to the user as illustrated.
- FIG. 15 illustrates an exemplary process flow to effect generating a call from the contacts menu. It is to be appreciated that a call can also be generated in accordance with a meeting request/appointment or a received communication.
- the system can present user information at 1502 .
- the information presented at 1502 can effect any desired task.
- the user can call or send a message to the contact by pressing “2” or “3” respectively. If the user chooses to call, the process continues to generate the call as illustrated.
- Illustrated in FIG. 16 is a process flow diagram to record a message.
- the message is recorded at 1602 .
- the user is presented with a variety of options to manage the recording. For example, the user can mark the message with high priority by depressing “1,1” or restrict access by depressing “5”. Once complete, the user can exit the system at 1606 .
- FIGS. 17 to 21 illustrate exemplary process flow diagrams of a UM system that employs a speech UI.
- a user can selectively employ disparate navigational mechanisms with respect to the application state machine (e.g., UM system).
- the process flow diagrams of FIGS. 17 to 21 illustrate the co-existence of the DTMF UI ( FIGS. 12 to 16 ) and a speech activated UI. It is to be understood that a user can employ any available UI at any time in connection with navigating within an application state machine.
- FIG. 17 an exemplary process task flow diagram of calling into a UM system is shown.
- the process illustrates a flow of procedures in accordance to a call initiated by an external caller 1202 and an internal caller 1204 .
- the process flow of the exemplary aspect is dependent upon the type of caller (e.g., external 1202 or internal 1204 ) as well as the contact number dialed at 1206 .
- An external caller will be prompted with the company information at 1208 whereby an internal caller with internal information at 1210 .
- a caller can contact a user located on the system by employing the directory search ( 1212 of FIG. 12 ) or by connecting directly to the extension ( 1214 ) by entering the numeric extension.
- the user can speak commands at 1702 to effect a desired option.
- the user can verbally instruct the system of a desired target person to contact.
- a user can speak the word “help” ( 1706 ) to access available system options.
- FIG. 18 illustrates a continued procedure flow diagram that includes voice options once access is granted.
- the user can be prompted with a main menu at 1302 .
- the user can be prompted to press “1” for voicemail, “2” for calendar options, “3” for contacts, “4” for email and so on.
- a user can choose “5” to send a message (e.g., voicemail, email) or “7” to set an OOF.
- Additional exemplary DTMF menu choices are illustrated in block 1302 .
- voice commands can be employed at 1802 in order to effect navigation.
- voicemail or email the system proceeds to 1304 whereby content can be presented to the user.
- the user can be presented with voicemail, email and/or meeting request content.
- logic and reasoning mechanisms can be employed to manage, organize, filter and/or sort the content as described supra.
- the user can be presented with an “after message” menu thereby presenting additional options.
- by pressing “4” a user can send a facsimile of a desired message.
- the user can opt to send a facsimile of the message to a desired location.
- voice commands shown at 1804 can be employed “during” or “after message” playback. In other words, the voice commands can be reused in either scenario. These reusable commands are listed on FIG. 18 as “anytime” commands.
- a list of exemplary DTMF play commands which are presented during playback of a message, are shown. Again, this list is not intended to be exhaustive in view of the functionality described herein. As illustrated, it will be appreciated that the system can effect a pause, undelete, repeat, etc., task in accordance with options shown at 1308 . In one aspect at 1806 , speech commands and options can be employed in connection with pausing a playback.
- the system can play the number of pending meetings.
- the user can manage the meeting appointments and/or requests by selecting any of various options. For example, the user can select option “2” to call the meeting location, “2,2” to call the meeting organizer or “3” to indicate a late arrival to the subject meeting.
- additional navigational options can be presented to the user as illustrated. For example, as illustrated, a user can employ speech commands at 1902 to effect calendar playback.
- FIG. 20 illustrates an exemplary process flow to effect generating a call from the contacts menu. It is to be appreciated that a call can also be generated in accordance with a meeting request/appointment or a received communication.
- the system can present user information at 1502 .
- the information presented at 1502 can effect any desired task.
- the user can call or send a message to the contact by pressing “2” or “3” respectively. If the user chooses to call, the process continues to generate the call as illustrated.
- the user can effect communication with a contact by employing speech commands. For example, the user can audibly instruct the system to “find” a particular person. Next, the user can audibly tell the system to send an email message to the located contact.
- Illustrated in FIG. 21 is a process flow diagram to record a message.
- the message is recorded at 1602 .
- the user is presented with a variety of options to manage the recording. For example, the user can mark the message with high priority by depressing “1,1” or restrict access by depressing “5”. Once complete, the user can exit the system at 1606 .
- a user at any time can opt to communicate with the system via another available UI (e.g., speech recognition).
- the user can speak the commands identified at 2102 thus navigating through the application state machine (e.g., UM system).
- the application state machine e.g., UM system
- FIG. 22 there is illustrated a block diagram of a computer operable to execute the disclosed architecture.
- FIG. 22 and the following discussion are intended to provide a brief, general description of a suitable computing environment 2200 in which the various aspects of the subject invention can be implemented. While the invention has been described above in the general context of computer-executable instructions that may run on one or more computers, those skilled in the art will recognize that the invention also can be implemented in combination with other program modules and/or as a combination of hardware and software.
- program modules include routines, programs, components, data structures, etc., that perform particular tasks or implement particular abstract data types.
- inventive methods can be practiced with other computer system configurations, including single-processor or multiprocessor computer systems, minicomputers, mainframe computers, as well as personal computers, hand-held computing devices, microprocessor-based or programmable consumer electronics, and the like, each of which can be operatively coupled to one or more associated devices.
- the illustrated aspects of the invention may also be practiced in distributed computing environments where certain tasks are performed by remote processing devices that are linked through a communications network.
- program modules can be located in both local and remote memory storage devices.
- a computer typically includes a variety of computer-readable media.
- Computer-readable media can be any available media that can be accessed by the computer and includes both volatile and nonvolatile media, removable and non-removable media.
- Computer readable media can comprise computer storage media and communication media.
- Computer storage media includes both volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer readable instructions, data structures, program modules or other data.
- Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital video disk (DVD) or other optical disk storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed by the computer.
- Communication media typically embodies computer-readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism, and includes any information delivery media.
- modulated data signal means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal.
- communication media includes wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared and other wireless media. Combinations of the any of the above should also be included within the scope of computer-readable media.
- FIG. 22 there is illustrated an exemplary environment 2200 for implementing various aspects of the invention that includes a computer 2202 , the computer 2202 including a processing unit 2204 , a system memory 2206 and a system bus 2208 .
- the system bus 2208 couples system components including, but not limited to, the system memory 2206 to the processing unit 2204 .
- the processing unit 2204 can be any of various commercially available processors. Dual microprocessors and other multi-processor architectures may also be employed as the processing unit 2204 .
- the system bus 2208 can be any of several types of bus structure that may further interconnect to a memory bus (with or without a memory controller), a peripheral bus, and a local bus using any of a variety of commercially available bus architectures.
- the system memory 2206 includes read only memory (ROM) 2210 and random access memory (RAM) 2212 .
- ROM read only memory
- RAM random access memory
- a basic input/output system (BIOS) is stored in a non-volatile memory 2210 such as ROM, EPROM, EEPROM, which BIOS contains the basic routines that help to transfer information between elements within the computer 2202 , such as during start-up.
- the RAM 2212 can also include a high-speed RAM such as static RAM for caching data.
- the computer 2202 further includes an internal hard disk drive (HDD) 2214 (e.g., EIDE, SATA), which internal hard disk drive 2214 may also be configured for external use in a suitable chassis (not shown), a magnetic floppy disk drive (FDD) 2216 , (e.g., to read from or write to a removable diskette 2218 ) and an optical disk drive 2220 , (e.g., reading a CD-ROM disk 2222 or, to read from or write to other high capacity optical media such as the DVD).
- the hard disk drive 2214 , magnetic disk drive 2216 and optical disk drive 2220 can be connected to the system bus 2208 by a hard disk drive interface 2224 , a magnetic disk drive interface 2226 and an optical drive interface 2228 , respectively.
- the interface 2224 for external drive implementations includes at least one or both of Universal Serial Bus (USB) and IEEE 1394 interface technologies.
- the drives and their associated computer-readable media provide nonvolatile storage of data, data structures, computer-executable instructions, and so forth.
- the drives and media accommodate the storage of any data in a suitable digital format.
- computer-readable media refers to a HDD, a removable magnetic diskette, and a removable optical media such as a CD or DVD, it should be appreciated by those skilled in the art that other types of media which are readable by a computer, such as zip drives, magnetic cassettes, flash memory cards, cartridges, and the like, may also be used in the exemplary operating environment, and further, that any such media may contain computer-executable instructions for performing the methods of the subject invention.
- a number of program modules can be stored in the drives and RAM 2212 , including an operating system 2230 , one or more application programs 2232 , other program modules 2234 and program data 2236 . All or portions of the operating system, applications, modules, and/or data can also be cached in the RAM 2212 . It is appreciated that the subject invention can be implemented with various commercially available operating systems or combinations of operating systems.
- a user can enter commands and information into the computer 2202 through one or more wired/wireless input devices, e.g., a keyboard 2238 and a pointing device, such as a mouse 2240 .
- Other input devices may include a microphone, an IR remote control, a joystick, a game pad, a stylus pen, touch screen, or the like.
- These and other input devices are often connected to the processing unit 2204 through an input device interface 2242 that is coupled to the system bus 2208 , but can be connected by other interfaces, such as a parallel port, an IEEE 1394 serial port, a game port, a USB port, an IR interface, etc.
- a monitor 2244 or other type of display device is also connected to the system bus 2208 via an interface, such as a video adapter 2246 .
- a computer typically includes other peripheral output devices (not shown), such as speakers, printers, etc.
- the computer 2202 may operate in a networked environment using logical connections via wired and/or wireless communications to one or more remote computers, such as a remote computer(s) 2248 .
- the remote computer(s) 2248 can be a workstation, a server computer, a router, a personal computer, portable computer, microprocessor-based entertainment appliance, a peer device or other common network node, and typically includes many or all of the elements described relative to the computer 2202 , although, for purposes of brevity, only a memory storage device 2250 is illustrated.
- the logical connections depicted include wired/wireless connectivity to a local area network (LAN) 2252 and/or larger networks, e.g., a wide area network (WAN) 2254 .
- LAN and WAN networking environments are commonplace in offices, and companies, and facilitate enterprise-wide computer networks, such as intranets, all of which may connect to a global communication network, e.g., the Internet.
- the computer 2202 When used in a LAN networking environment, the computer 2202 is connected to the local network 2252 through a wired and/or wireless communication network interface or adapter 2256 .
- the adaptor 2256 may facilitate wired or wireless communication to the LAN 2252 , which may also include a wireless access point disposed thereon for communicating with the wireless adaptor 2256 .
- the computer 2202 can include a modem 2258 , or is connected to a communications server on the WAN 2254 , or has other means for establishing communications over the WAN 2254 , such as by way of the Internet.
- the modem 2258 which can be internal or external and a wired or wireless device, is connected to the system bus 2208 via the serial port interface 2242 .
- program modules depicted relative to the computer 2202 can be stored in the remote memory/storage device 2250 . It will be appreciated that the network connections shown are exemplary and other means of establishing a communications link between the computers can be used.
- the computer 2202 is operable to communicate with any wireless devices or entities operatively disposed in wireless communication, e.g., a printer, scanner, desktop and/or portable computer, portable data assistant, communications satellite, any piece of equipment or location associated with a wirelessly detectable tag (e.g., a kiosk, news stand, restroom), and telephone.
- any wireless devices or entities operatively disposed in wireless communication e.g., a printer, scanner, desktop and/or portable computer, portable data assistant, communications satellite, any piece of equipment or location associated with a wirelessly detectable tag (e.g., a kiosk, news stand, restroom), and telephone.
- the communication can be a predefined structure as with conventional network or simply an ad hoc communication between at least two devices.
- Wi-Fi Wireless Fidelity
- Wi-Fi is a wireless technology like a cell phone that enables such devices, e.g., computers, to send and receive data indoors and out; anywhere within the range of a base station.
- Wi-Fi networks use radio technologies called IEEE 802.11 (a, b, g, etc.) to provide secure, reliable, fast wireless connectivity.
- IEEE 802.11 a, b, g, etc.
- a Wi-Fi network can be used to connect computers to each other, to the Internet, and to wired networks (which use IEEE 802.3 or Ethernet).
- Wi-Fi networks operate in the unlicensed 2.4 and 5 GHz radio bands, at an 11 Mbps (802.11 a) or 54 Mbps (802.11 b) data rate, for example, or with products that contain both bands (dual band), so the networks can provide real-world performance similar to the basic 10 BaseT wired Ethernet networks used in many offices.
- the system 2300 includes one or more client(s) 2302 .
- the client(s) 2302 can be hardware and/or software (e.g., threads, processes, computing devices).
- the client(s) 2302 can house cookie(s) and/or associated contextual information by employing the subject invention, for example.
- the system 2300 also includes one or more server(s) 2304 .
- the server(s) 2304 can also be hardware and/or software (e.g., threads, processes, computing devices).
- the servers 2304 can house threads to perform transformations by employing the subject invention, for example.
- One possible communication between a client 2302 and a server 2304 can be in the form of a data packet adapted to be transmitted between two or more computer processes.
- the data packet may include a cookie and/or associated contextual information, for example.
- the system 2300 includes a communication framework 2306 (e.g., a global communication network such as the Internet) that can be employed to facilitate communications between the client(s) 2302 and the server(s) 2304 .
- a communication framework 2306 e.g., a global communication network such as the Internet
- Communications can be facilitated via a wired (including optical fiber) and/or wireless technology.
- the client(s) 2302 are operatively connected to one or more client data store(s) 2308 that can be employed to store information local to the client(s) 2302 (e.g., cookie(s) and/or associated contextual information).
- the server(s) 2304 are operatively connected to one or more server data store(s) 2310 that can be employed to store information local to the servers 2304 .
Landscapes
- Engineering & Computer Science (AREA)
- Signal Processing (AREA)
- Food Science & Technology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Multimedia (AREA)
- Telephonic Communication Services (AREA)
- Information Transfer Between Computers (AREA)
Abstract
A system that concurrently provides multiple user interface (UI) mechanisms that facilitate control of an application state machine (e.g., unified message system). More particularly, the invention can create two relatively distinct user experiences, one via dual tone multi-frequency (DTMF) navigation and another through speech recognition navigation of a unified message system. In accordance therewith, one single underlying state machine can be used. Navigation and flow control (e.g., state transitions) in the state machine can be leveraged by multiple UI mechanisms that actively co-exist. The invention introduces speech recognition features together with other input mechanisms to drive the UI of an application state machine (e.g., unified messaging system). The speech recognition UI can be designed to provide a natural navigation through the application independent of a DTMF UI.
Description
- This application is a Continuation of co-pending U.S. application Ser. No. 11/086,828 entitled “Selectable State Machine User Interface System” filed Mar. 22, 2005, which claims priority to U.S. Provisional Application No. 60/620,826 filed Oct. 20, 2004 entitled “Unified Messaging System”, which are incorporated herein by reference.
- This invention is related to messaging systems and more particularly to a selectable user interface system that can employ variable interactive mechanisms to navigate through and control an underlying application state machine.
- A state machine is a computing device designed with the operational states required to perform a particular task or to solve a specific problem. There are countless special-purpose devices built as state machines. A voicemail system is an example of such an application state machine.
- Today, many state machines, and particularly voicemail systems, employ a dual tone multi-frequency (DTMF) navigational technique that enables a user to control the state machine. In other words, a user can navigate through voicemail menus and options by pressing appropriate keys on a telephone keypad. Many of these DTMF navigational techniques require a user to memorize option choices that correspond to the numbers on a keypad.
- Other DTMF systems employ a navigational technique that allows a user to spell words via a telephone keypad. This is often cumbersome and very time consuming. Other variations of DTMF voicemail systems employ voice media processing components that present to a user a set of available options. For example, a user is frequently addressed with a specific prompt such as, “please enter your password” or “press one to review all new messages.” While these audible prompts are sometimes useful, they tend to be time consuming and, nonetheless, often still require the user to employ DTMF techniques via a keypad.
- Recent efforts have been directed to utilizing speech recognition techniques to enable a user to navigate through state machine menus. These efforts have been very limited in that they only recognize preprogrammed menu selections and/or data. For example, a system can be programmed to recognize the names of individuals employed by a company.
- Despite which technique is provided, state machines today are limited to employing a single navigational technique at any given time. In other words, in accordance with conventional systems, a user initially chooses to navigate through a state machine via DTMF or voice command techniques. However, with respect to these systems, both options are usually not simultaneously or concurrently available and the user does not have the option of toggling between the navigational systems during any one session.
- Although attempts have been made to employ either DTMF or voice recognition techniques to effect navigation through a state machine (e.g., a voicemail system), a substantial need exists for a system that simultaneously or concurrently provides the functionality of multiple disparate navigational techniques thereby permitting a user to select a preferred user interface technique at any time throughout a session. Additionally, a need exists for a system and/or methodology that employs voice recognition mechanisms to effect versatility in connection with state machine navigation.
- The following presents a simplified summary of the invention in order to provide a basic understanding of some aspects of the invention. This summary is not an extensive overview of the invention. It is not intended to identify key/critical elements of the invention or to delineate the scope of the invention. Its sole purpose is to present some concepts of the invention in a simplified form as a prelude to the more detailed description that is presented later.
- The subject invention, in one aspect thereof, allows a user to connect to a server (e.g., messaging server) via a telephonic communication component thereby accessing the rich information therein. The invention comprises a system and/or methodology that unifies a telephonic communication component and/or system with an application state machine (e.g., messaging server component and/or system). More particularly, in one aspect, the invention creates two relatively distinct user experiences, one through the use of dual tone multi-frequency (DTMF) navigation and one through speech recognition navigation. In accordance thereto, one single underlying state machine can be used.
- Navigation and flow control (e.g., state transitions) in the state machine can be leveraged by multiple user interface (UI) mechanisms that co-exist actively. The invention introduces speech recognition features together with other input mechanisms to drive the user interface of an application state machine (e.g., a unified messaging system (UM)). In accordance with an aspect of the invention, the speech recognition UI is designed to provide a natural navigation through the application independent of a DTMF user interface. The user can navigate through the speech menu without having to enter any DTMF data.
- In accordance with the subject invention, both the DTMF-based menu infrastructure and the speech recognition menu structure can be active and available to the user at any time. For example, at any time during the experience the user can switch from speech mode to DTMF driven mode. Because the switch does not need to be explicit, the user can issue either a DTMF or a speech command at any time.
- In another aspect, the prompts that will be played for the user upon entering any part of the application can be based on the UI mechanism that the user used to enter the state. For example, the speech menu will be played if speech commands brought the user to the state. On the other hand, the DTMF menu will be played if a DTMF command brought the user to the state.
- In yet another aspect, from either the speech or DTMF UI the user can, at any time, say “help” to review available speech commands or press “0” to review available DTMF commands. In accordance with the co-existence of the UI mechanisms, both sets of commands are available at all times. The default UI that is presented to the user on first entry to the UM system can be set as a personal option. Rule-based logic and/or artificial intelligence reasoning mechanisms can be employed to determine the default UI.
- The underlying application state machine is leveraged for the multiple co-existent UI models. Specific to aspects of the speech UI are additional design paradigms selected to exploit some unique features of speech recognition. For example, the speech UI can have distinct prompts and menus for handling user input, input validation, and error handling for low confidence recognition results.
- In still other aspects, a set of “anytime” shortcuts can be available at all times to jump to the main functionality of the underlying state machine (e.g., unified messaging application). In the unified messaging application example, these main functionality locations can be voicemail, calendar, call someone, email, and personal options.
- Each speech menu can also have menu commands that are specific to the menu context. For example, with access to an application programming interface, the invention can add functionality to enable telephonic generation and retrieval of email communications. The invention can further facilitate accepting or canceling calendar items (e.g., meeting requests and appointments) via a telephonic communications device.
- To the accomplishment of the foregoing and related ends, certain illustrative aspects of the invention are described herein in connection with the following description and the annexed drawings. These aspects are indicative, however, of but a few of the various ways in which the principles of the invention can be employed and the subject invention is intended to include all such aspects and their equivalents. Other advantages and novel features of the invention will become apparent from the following detailed description of the invention when considered in conjunction with the drawings.
-
FIG. 1 is a high-level exemplary system architecture that facilitates state machine user interface selection in accordance with an aspect of the invention. -
FIG. 2 is an exemplary system that illustrates a user interface store having disparate state machine user interface components in accordance with a disclosed aspect. -
FIG. 3 illustrates a system architecture that facilitates user interface selection with respect to a unified messaging system in accordance with a disclosed aspect. -
FIG. 4 illustrates a user interface selection component that employs a rule-based logic engine in accordance with a disclosed aspect. -
FIG. 5 illustrates a user interface selection component that employs an artificial intelligence component in accordance with an exemplary aspect. -
FIG. 6 illustrates a high-level exemplary system architecture that facilitates communication in accordance with an aspect of the subject invention. -
FIG. 7 illustrates an exemplary system architecture that employs a unified messaging system in accordance with a disclosed aspect. -
FIG. 8 is a call control component that facilitates content conversion and interpretation in accordance with a disclosed aspect. -
FIG. 9 illustrates a flow diagram of an exemplary communication methodology in accordance with a disclosed aspect. -
FIG. 10 illustrates a call control component that employs a rule-based logic engine in accordance with a disclosed aspect. -
FIG. 11 illustrates a call control component that employs an artificial intelligence component in accordance with an aspect. -
FIG. 12 is an exemplary flow diagram of calling into a unified messaging system in accordance with a disclosed aspect. -
FIG. 13 is an exemplary flow diagram of accessing voicemail and email in accordance with an aspect. -
FIG. 14 is an exemplary flow diagram of accessing an electronic calendar in accordance with an exemplary aspect. -
FIG. 15 is an exemplary flow diagram of accessing electronic contacts in accordance with a disclosed aspect. -
FIG. 16 is an exemplary flow diagram of recording a message in accordance with a disclosed aspect. -
FIG. 17 is an exemplary flow diagram of calling into a unified messaging system that employs voice commands in accordance with a disclosed aspect. -
FIG. 18 is an exemplary flow diagram of accessing voicemail and email that employs voice commands in accordance with an aspect of the invention. -
FIG. 19 is an exemplary flow diagram of accessing an electronic calendar that employs voice commands in accordance with a disclosed aspect. -
FIG. 20 is an exemplary flow diagram of accessing electronic contacts that employs voice commands in accordance with a disclosed aspect. -
FIG. 21 is an exemplary flow diagram of recording a message that employs voice commands in accordance with a disclosed aspect. -
FIG. 22 illustrates a block diagram of a computer that can execute the disclosed architecture. -
FIG. 23 illustrates a schematic block diagram of an exemplary computing environment in accordance with the subject invention. - The subject invention is now described with reference to the drawings, wherein like reference numerals are used to refer to like elements throughout. In the following description, for purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the subject invention. It may be evident, however, that the subject invention can be practiced without these specific details. In other instances, well-known structures and devices are shown in block diagram form in order to facilitate describing the subject invention.
- As used in this application, the terms “component” and “system” are intended to refer to a computer-related entity, either hardware, a combination of hardware and software, software, or software in execution. For example, a component can be, but is not limited to being, a process running on a processor, a processor, an object, an executable, a thread of execution, a program, and/or a computer. By way of illustration, both an application running on a server and the server can be a component. One or more components can reside within a process and/or thread of execution, and a component can be localized on one computer and/or distributed between two or more computers.
- As used herein, the term to “infer” or “inference” refer generally to the process of reasoning about or inferring states of the system, environment, and/or user from a set of observations as captured via events and/or data. Inference can be employed to identify a specific context or action, or can generate a probability distribution over states, for example. The inference can be probabilistic—that is, the computation of a probability distribution over states of interest based on a consideration of data and events. Inference can also refer to techniques employed for composing higher-level events from a set of events and/or data. Such inference results in the construction of new events or actions from a set of observed events and/or stored event data, whether or not the events are correlated in close temporal proximity, and whether the events and data come from one or several event and data sources.
- Referring now to
FIG. 1 , there is illustrated a high-level architectural diagram of asystem 100 in accordance with an aspect of the subject invention. Generally, thesystem 100 can include a user interface (UI)selection component 102, aUI store 104 and astate machine component 106. It will be understood and appreciated that a novel aspect of the invention is to decouple the UI mechanisms from thestate machine 106. The invention can employ multiple UI mechanisms to effect navigation within an application state machine. Further, a novel aspect of the invention can concurrently employ multiple UI interfaces (e.g., dual tone multi-frequency (DTMF), speech recognition). - In other words, in accordance with novel aspects of the invention, multiple UI mechanisms can be active at any time allowing a user to freely switch between UI mechanisms. In the example that follows, this invention employs a novel design that allows for the co-existence of two or more UI mechanisms (e.g., DTMF and speech recognition) that enable telephone access to a user inbox. Furthermore, the two or more UI mechanisms can simultaneously leverage the underlying application state machine.
- The exemplary aspects that follow are directed to the use of the
UI selection component 102 and theUI store 104 in connection with a voicemail application (e.g., state machine 106). However, it is to be appreciated that thestate machine component 106 can be any application state machine. Accordingly, theUI selection component 102 can effect utilization of any available UI mechanisms. By way of example, in operation, a user can employ theUI selection component 102 to select a DTMF navigational technique to control the operation of an underlying state machine (e.g., voicemail system). As well, theUI selection component 102 can be employed to facilitate a speech recognition navigational technique. It is to be understood that theUI selection component 102 can be continuously available thereby providing a user the ability to toggle between UI navigational techniques at any time during a session. - While the exemplary aspects described herein are directed to systems that employ a DTMF and/or voice recognition navigational technique, it is to be understood that any navigational technique can be employed without departing from the spirit and/or scope of the functionality of the invention. For example, other aspects can employ a pointing device to effect navigation with reference to a state machine.
- Turning now to
FIG. 2 , there is illustrated a more detailed architectural diagram of an aspect of the invention. As described supra, in order to create two distinct user experiences, one single underlyingstate machine component 106 can be used in connection with multiple UI mechanisms. In accordance therewith, flow control (e.g., state transitions) in the state machine can be leveraged by both UI mechanisms. - As shown in
FIG. 2 , theUI store 104 can include 1 to N UI components, where N is an integer. It will be appreciated that 1 to N UI components can be referred to individually or collectively asUI components 202. A more specific example of theUI store 104 is illustrated inFIG. 3 . - Referring to the exemplary aspect of
FIG. 3 , as illustrated, theUI store 104 can includeUI components 202 including, but not limited to, a DTMF UI component and a speech UI component. Additionally, thestate machine component 106 can include aunified messaging system 302. Theunified messaging system 302 can unite a telephonic system with an electronic messaging and/or organization system. An exemplary unified messaging system will be discussed in detail infra. - As described supra, the invention introduces speech recognition features together with DTMF features to drive the UI of
unified messaging system 302. In accordance with an aspect, thespeech recognition UI 202 can be designed to provide a natural navigation through the application (e.g., state machine 106) independent of theDTMF UI 202. Because the invention decouples these UI mechanisms, the user will be able to drive through the speech menu without having to enter any DTMF input(s). - It is a novel feature of the invention to enable both the DTMF-based menu infrastructure and the speech recognition menu structure to be active and available to the user at any time. At any time during a session or experience, the user can independently switch from speech mode to DTMF driven mode. The switch need not be explicit; the user need only issue either a DTMF signal or a speech command to automatically effect the transition.
- Continuing with the voicemail system example, the prompts that will be played for the user on entering any part of the application will be based on the UI mechanism the user employed to enter the state. If speech commands were employed to bring the user to a state, the speech menu will be played. On the other hand, if a DTMF command brought the user to the state, the DTMF menu will be played. Additionally, from either the speech or
DTMF UI 202 the user can employ “help” to determine the commands. For example, at any time the user can say “help” to review available speech commands or press “0” to review available DTMF commands. Both sets of commands are available at all times. - A default UI that is presented to the user on first entry to the
unified messaging system 302 can be set as a personal option. As well, it will be appreciated that rule-based logic and/or artificial intelligence (AI) techniques can be employed to automate the selection process. These rule-based logic and AI based reasoning techniques will be discussed in greater detail infra. - Specific to the
speech UI 202 are additional design paradigms selected to exploit some novel features of speech recognition. Thespeech UI 202 can have distinct prompts and menus for handling user input, input validation, and error handling for low confidence recognition results. Further, the DTMF prompts will not be reused or converted to “press or say”. The speech prompts and menus can go through a full usability pass including user education review. Menus can be generated to sound natural for the user. For example, a menu can be generated to say “Would you like to make a call or access your mailbox?” rather than “Say ‘make a call’ to call someone or ‘access mailbox’ to access your mailbox”. - A set of “anytime” shortcuts can be available at all times to jump to the main functionality of the
unified messaging application 302. As will be discussed in detail infra, the main functionality can include voicemail, calendar, call someone, email, and personal options. Each speech menu can also have menu commands that are specific to the menu context. - In an alternative aspect, the invention can limit access to features of the
unified messaging application 106. In other words, if it makes the application flow more natural, some more obscure or less important features may only be provided through DTMF mode. These features can be preprogrammed and/or defined by a user as desired. As will be better understood with reference toFIGS. 17 to 21 , within the speech menus, menu flexibility can be increased since there is no concern around reusing DTMF digits. For example, in the voicemail, email and calendar playback menus there does not need to be a distinction between the “during playback” and the “after message” menu. - It will be appreciated that, in connection with using speech commands, the DTMF paths of typing an extension and spelling out a name collapse to a single item with “Find [Person Name]”. Furthermore, the infrastructure around contact navigation can also collapse into “Find [Person Name].” Occasional word combinations such as “I'll be 15 minutes late” combine what was previously two entries under DTMF mode (e.g., an entry for late and then a separate entry for the time.)
- An overview of the speech recognition menu structure is described with reference to
FIGS. 17 to 21 . These figures display both the DTMF menu infrastructure and the speech user interface overlay. The speech diagrams are intended to show the main grammar elements of an exemplary UI and the overall flow rather than every specific prompt and error handling condition. The flow diagrams included are not intended to limit the scope and/or novel functionality of the invention in any way. - The aforementioned novel aspects of the invention can employ rule-based logic and/or AI reasoning technology to effect, predict and/or infer an action. More particularly, it will be appreciated that the subject invention (e.g., in connection with state analysis, UI selection, voice recognition) can employ various rule-based and/or AI based schemes for carrying out various aspects thereof.
- Continuing with the above-described scenarios and with reference to
FIG. 4 , a block schematic view of an alternativeUI selection component 102 is shown. As illustrated,UI selection component 102 can include a rule-basedlogic engine 402. As will later be described with reference toFIG. 5 , an optional AI component (not shown) can be used together with, or in place of, the rule-basedlogic engine 402 to automatically infer an action or set of actions to be employed in connection with the functionality of theUI selection component 102 described supra. - In the exemplary aspect of
FIG. 4 , the rule-basedlogic engine 402 can be programmed or configured in accordance with a predefined preference (e.g., a rule). For example, a rule (e.g., logic) can be constructed to automatically select a specific UI based upon defined criteria. Accordingly, the system can automatically select and employ a specific UI. More particularly, a rule can be established to take into consideration the origination location (e.g., internal, external, Internet . . . ), originator, target location, etc., to automatically select and employ a UI. - A schematic diagram of another alternative aspect of the subject invention is illustrated in
FIG. 5 . TheUI selection component 102 of this alternative aspect illustrated inFIG. 5 employs anAI component 502 that can automatically infer and/or predict an action. This alternative aspect can optionally include an inference module (not shown) that facilitates automatic selection of theUI selection component 102. - In accordance with this aspect, the
optional AI component 502 can facilitate automatically performing various aspects (e.g., state analysis, UI selection, and voice recognition) of the subject invention as described herein. TheAI component 502 can optionally include an inference component (not shown) that can further enhance automated aspects of the AI component utilizing, in part, inference-based schemes to facilitate inferring intended actions to be performed at a given time and/or state. The AI-based aspects of the invention can be effected via any suitable machine-learning based technique and/or statistical-based techniques and/or probabilistic-based techniques. - In the alternate aspect, as further illustrated by
FIG. 5 , the subject invention (e.g., in connection with state analysis, UI selection, and voice recognition) can optionally employ various AI-based schemes for automatically carrying out various aspects thereof. Specifically, theAI component 502 can optionally be provided to implement aspects of the subject invention based upon AI processes (e.g., confidence, inference . . . ). For example, a process for initiating a UI based upon user preferences or sender identification can be facilitated via an automatic classifier system and process. Further, theoptional AI component 502 can be employed to facilitate an automated process of selecting a UI based upon a predicted and/or inferred preference. - A classifier is a function that maps an input attribute vector, x=(x1, x2, x3, x4, xn), to a confidence that the input belongs to a class, that is, f(x)=confidence(class). Such classification can employ a probabilistic and/or statistical-based analysis (e.g., factoring into the analysis utilities and costs) to prognose or infer an action that a user desires to be automatically performed.
- A support vector machine (SVM) is an example of a classifier that can be employed. The SVM operates by finding a hypersurface in the space of possible inputs, which hypersurface attempts to split the triggering criteria from the non-triggering events. Intuitively, this makes the classification correct for testing data that is near, but not identical to training data. Other directed and undirected model classification approaches include, e.g., naive Bayes, Bayesian networks, decision trees, and probabilistic classification models providing different patterns of independence can be employed. Classification as used herein also is inclusive of statistical regression that is utilized to develop models of priority.
- As will be readily appreciated from the subject specification, the subject invention can employ classifiers that are explicitly trained (e.g., via a generic training data) as well as implicitly trained (e.g., via observing user behavior, receiving extrinsic information). For example, SVM's can be configured via a learning or training phase within a classifier constructor and feature selection module. In other words, the use of expert systems, fuzzy logic, support vector machines, greedy search algorithms, rule-based systems, Bayesian models (e.g., Bayesian networks), neural networks, other non-linear training techniques, data fusion, utility-based analytical systems, systems employing Bayesian models, etc. are contemplated and are intended to fall within the scope of the hereto appended claims.
- Referring now to
FIG. 6 , there is illustrated a high-level architectural diagram of asystem 600 in accordance with an aspect of the subject invention. Generally, thesystem 600 can include a unified messaging (UM)system component 602, atelephonic communication component 604 and a computer application, state machine and/orserver component 606. As described supra, the UI selection component 102 (FIG. 1 ) and subsequent functionality of the subject invention can be employed in connection with theunified messaging system 602. More particularly, as illustrated inFIG. 6 , the subject invention can be employed to enable a UI to be selected thus effecting telephonic access to rich data maintained within a server component. This telephonic access can be effected via aunified messaging system 602. - The
telephonic communication component 604 can include any device capable of communicating voice and/or audible sounds and signals (e.g., DTMF). By way of example, thetelephonic communication component 604 can include a traditional wired telephone as well as wireless (e.g., cellular) telephones. Further, thetelephonic communication component 604 can be effected via a portable wireless device and/or computer that employs voice over Internet protocol (VOIP) or fax over Internet protocol (FOIP). - The
server component 606 can be employed to host any communication and/or organizational application. For example, theserver component 606 can host communication applications including, but not limited to, voicemail, email, text messaging applications or the like in connection with aspects of the subject invention. - In one aspect, the
UM system 602 can unite thetelephonic communications system 604 with the server component 606 (e.g., mailbox server). Accordingly, thetelephonic communications system 604 can access the rich information maintained in theserver component 606 thereby, unifying the systems. AlthoughFIG. 6 illustrates a singletelephonic communications system 604 and asingle server component 606, it is to be understood and appreciated that the invention is extensible thereby being capable of uniting multiple telephonic and/or server systems. - Referring now to
FIG. 7 , there is illustrated an exemplary system architecture in accordance with an aspect of the invention. Thesystem 700 generally includes theUM system 602,telephonic communications system 604 andserver component 606. In one aspect, as illustrated, thetelephonic communication component 604, theserver component 606 and a plurality ofclients 702 can be connected to theUM system 602 via a wired local area network (LAN) connection 704 (e.g., Ethernet). Although three specific clients 702 (e.g., smart phone, desktop, laptop) are shown, it is to be appreciated that any number of clients 702 (e.g., email clients) can be employed without departing from the spirit and/or scope of the subject invention. It will be understood that, in alternate aspects, connections illustrated inFIG. 7 can be wired, wireless or a combination thereof. - Furthermore, it is to be appreciated that the
clients 702 can employ disparate communication techniques in order to communicate with theserver component 606. For example, a smart phone, pocket personal computer (PC) or other handheld device can employ specialized communication software to synchronize to themailbox server 706 or theactive directory server 708. In another example, a desktop (e.g., client 702) can employ a wired (e.g., Ethernet) connection thereby gaining access to theserver component 606. Yet another exemplary client of the three clients 702 (e.g., laptop) can employ a wireless Internet connection to access theserver component 606. - Turning now to the
UM system component 602, acall control component 710, text-to-speech (TTS)engine 712,speech recognition engine 714, voicemedia processing component 716, faxmedia processing component 718 andUM application component 720 can be provided. Each of these components can be better understood with reference to the exemplary scenarios infra. - The
telephonic communications component 604 can include apublic telephone 722 connected via a public switched telephone network (PSTN) 724, a private (e.g., in-house company)phone 726 connected via a private branch exchange (PBX) 728 or private telephone switchboard, and computer system 730 that employs VOIP or FOIP viagateway 732. It will be appreciated that these components are exemplary and are not intended to be an exhaustive list of thetelephonic communications component 604. In other words,telephonic communications component 604 can include any mechanism capable of generating and/or transmitting audible (e.g., voice, fax) signals. - Additionally,
UM system 602 can include aUI selection component 734 to effect selection of an appropriate UI as described supra. Although not illustrated inFIG. 7 , it is to be understood that theUI selection component 734 can communicate with a UI store (not shown) in order to effect the desired and/or selected UI. It will further be appreciated that the UI store (not shown) can be co-located or located remotely from theUI selection component 734. For example, the UI store (not shown) can be located withinserver component 606. -
FIG. 8 illustrates an implementation of thecall control component 710. More particularly,UM application component 720 can include anemail integration component 802, acalendar integration component 804, a contactlist integration component 806 and asecurity component 808. Each of these components (802, 804, 806, 808) will be better understood upon a discussion of the exemplary scenarios that follow. Although theUM application component 720 ofFIG. 8 illustrates four exemplary functional components, it is to be appreciated that the functionality of these components described in the following scenarios can be combined (or separated) into disparate components. As well, it will be appreciated that the components described herein can be co-located or remotely located without departing from the scope of the subject invention. - In order to provide context to the invention, the following exemplary scenarios are provided. It is to be appreciated that the following scenarios are included merely to provide context to the invention. It will also be appreciated that additional scenarios exist which are not specifically disclosed herein. To this end, the scenarios infra are not intended to limit the scope and/or functionality of the subject invention.
- A first exemplary scenario is directed to a discussion of the
email integration component 802. More particularly, this scenario is directed to integration of an “out-of-office” (OOF) voicemail greeting and email. With reference again toFIG. 7 , conventionally, the OOF status can be separately set on both a user specific email account (e.g., via client 702) or voicemail account (e.g., viatelephone 722, 726). As will be appreciated, an OOF email auto-reply generated via the server component as well as the voicemail extended absence (e.g., OOF) greeting often contain similar information. The subject invention can facilitate setting both the email and the voicemail OOF status from a single location in a single operation. - With continued reference to
FIG. 7 , when a user sets the OOF status from the email application (e.g., client 702), theUM system 602 can integrate (e.g., synchronize) the OOF message into thetelephonic communications component 604. Accordingly, calls to the user via thetelephonic communications system 604 will include an audible version of the OOF extended absence greeting. Through the use of text-to-speech conversion techniques of theTTS engine 712, this greeting can audibly render the text contained in the email OOF auto-reply. - Similarly, in the event that a user sets the OOF from the telephonic communications component 604 (e.g., telephone 722), through voice recognition techniques of the
speech recognition engine 714, theunified messaging system 602 can automatically set an email OOF message in themailbox server 706. Thereafter, a sender of an email to the user will receive a text version of the OOF auto-reply that was originated via telephonic communication mechanisms. - A second scenario is directed to the
calendar integration component 804 and particularly to calendar or planner access and management. Today, users have limited access to their calendar from telephonic systems (e.g., 722, 726, 730) and can only read the information corresponding to appointments, which is maintained withinserver component 606. In accordance with an aspect of the subject invention, users can manage appointments and initiate communications related to appointments via the telephonic communications component 604 (e.g.,telephone speech recognition component 714 of theUM system 602. For example, via thetelephonic communications device 604, a user can accept, decline, modify, cancel, and communicate with respect to appointments contained within theserver component 606. - In one aspect, a telephonic session with the
UM system 602 can be automatically transferred to a telephone of a meeting location. In other words, via thetelephonic communications component 604, a user can connect to theUM system 602 whereby, based upon a calendar (e.g., appointment) entry in theserver component 606, a determination can be made with respect to an appropriate transfer target location for the call. Finally, by contacting theUM system 602 via anytelephonic communications device 604, a user can initiate a targeted or broadcast electronic message (e.g., email) to the meeting participants. For example, a user can inform meeting participants of a delayed arrival to a meeting. - Yet another aspect is directed to remote access of contact list entries via the contact
list integration component 806. In connection with systems today, users cannot telephonically access their personal contact list, which is frequently stored on the email system (e.g., server 606). In accordance with an aspect of the invention, from the telephonic communications device 604 (e.g.,telephone server 606. If desired, the user can immediately connect to or leave a message (e.g., voice, text, email) for the contact. If the contact is an external contact with only an email address, this invention can allow a user to communicate an audible message by creating an email with a sound file attachment (e.g., .wav) thereby forwarding the communication to the intended target. - Turning now to a discussion of the
security component 808 functionality of theUM application component 720, the subject invention can be employed to reset a personal identification number (PIN) via an email or other application. By way of example, when users forget their telephone voicemail PIN (e.g., password), they frequently have to contact an information technology (IT) helpdesk or system administrator to reset it. In accordance with theUM system 602, this invention can allow a user to log into their email on theserver component 606 using domain credentials and reset/change their telephone voicemail PIN (e.g., password). - Yet another scenario involving the
security component 808 is directed to digital rights management for voicemail. When leaving a voicemail today on the telephone, senders do not have any way to restrict the access to and/or management of the message. With this invention, voice messages can be sent with restricted access. For instance, a voicemail can be sent whereby a recipient can listen to the voicemail, but can be restricted from forwarding or saving the message. - Referring again to the
email integration component 802, conventionally, in situations where listening to a sound file is not appropriate (e.g., in a meeting), users are unable find out the contents of their voicemail. The UM system 602 (e.g., email integration component 802) of the subject invention provides for text preview of voicemails (and vice versa). In accordance with theUM system 602, via anemail client 702, users can textually access voicemails on their computer, laptop or other suitable device (e.g., smart phone). Through voice recognition techniques (e.g., speech recognition engine 714), the subject invention can convert the voicemail thereby transcribing it into text. Therefore, users can review the content of the voice message without playing the sound file. Similarly, theTTS engine 712 can be employed to convert and deliver email communications via thetelephonic communications device 604. - Additionally, the invention can facilitate converting electronic message attachments (e.g., word processing documents) to audible files (e.g., voice). This can be accomplished via the
TTS component 712. In accordance therewith, in addition to rendering the audible content of an electronic communication, the subject invention can render an audible file that contains the content of a textual attachment. It will be understood that the TTS conversion techniques and navigational concepts described herein can be applied to file system directories and the like. In other words, thespeech recognition engine 714 can be employed to effect navigation within a file system directory. Accordingly, theTTS engine 712 can be employed to convert a target document to speech thus rendering the audible (e.g., voice) file to a user via thetelephonic communications component 604. - Another aspect is directed to employing the
email integration component 802 to facilitate analyzing the content of an email communications message. For example, when email messages are audibly rendered to a user via thetelephonic communications device 604, (e.g., telephone), often the messages can be very long and difficult to follow. A novel feature of an aspect of this invention can employ logic and/or reasoning techniques to identify the most important messages and/or sentences (e.g., content) included in a communication. These logic and/or reasoning mechanisms can employ rule-based and/or artificial intelligence (AI) mechanisms to effect filtering and/or sorting of the content. Once identified, a streamlined or condensed version of the content can be audibly rendered to a user. - As previously described, the
UM system 602 can limit the audible playback of email messages to include only the most important messages and/or sentences included therein. It will be appreciated that because the content can be filtered and/or sorted in accordance with a desired algorithm (e.g., rule based, AI), the quantity and length of the message(s) that are read to the user can be decreased. Likewise, important voicemail messages can be handled (e.g., filtered, sorted) in the same manner. Of course, when people receive many email messages over the telephone it can be difficult to navigate to the particular messages that are important to the user. With this invention, important voicemails can be identified by theUM system 702 and read to the user first. As well, voice messages can be filtered whereby only portions of the voicemails are read to the user. - Another novel feature of the subject invention is directed to facsimile transmissions. By way of example, in some organizations, all inbound facsimiles are directed to one mailbox where a router looks at the facsimile and directs it, via electronic means, to the appropriate person in the organization. If the facsimile message contains sensitive information, the person who is redirecting the facsimile would have access to information that was intended to be confidential. In accordance with an aspect of this invention, the
UM system 602 can employ the faxmedia processing component 718 to separate the cover page from the remaining pages of a facsimile transmission. Therefore, the router who is redirecting the message can only have access to the cover page of the facsimile and not the confidential contents. - In another scenario, when people call a user and are transferred to the voicemail system, but do not leave a message, the callee does not always know that someone has called. In accordance with an aspect of this invention, the
UM system 602 can generate and forward an email message to the callee thereby notifying that someone has called but did not leave a voicemail message. - Still another novel aspect of the invention is directed to policy and compliance integration with an email system. Conventionally, voice messages are stored on a server (e.g., server component 606) and are retained and/or archived according to an arbitrary or preprogrammed policy. For example, one such policy can delete messages after a specific number of days (e.g., 21). With this invention (e.g., integration with the email system), a user can control which messages are deleted or archived from the
telephonic communications component 604. As well, logic and/or reasoning mechanisms can be applied to automatically manage retention and/or deletion of messages. - Yet another aspect of the invention is directed to a system and/or methodology to permit a user to select a keystroke profile scheme. It will be appreciated that different voicemail systems can have different keys to navigate and control the telephone user interface. If users change from one system to another, they often have to learn the new keys to use the new system. With this invention, button profiles can be employed for the most commonly used systems. Accordingly, a user can choose a predetermined profile that is familiar to that particular user. As well, buttons can be programmed as desired by a specific user. In this manner, the user does not necessarily have to learn a new keystroke profile scheme to access a new system.
- Another novel aspect of the
security component 808 is directed to employing voice recognition techniques to effect voicemail authentication. Today, when users access their mailbox from the telephone, they typically key in a PIN as their password. In many cases, this is not very secure and is inconvenient to the user (e.g., accessing while driving a car). With this invention, users can audibly speak a phrase or desired PIN to gain access to their mailbox. This can be more convenient and also more secure as the voice phrase must match a previously recorded phrase by the user. As well, the system can be programmed to learn qualities (e.g., voice tonal qualities) specific to a user. - Yet another exemplary aspect is directed to end-user configuration of menus and prompts. When end-users access their mailbox via the telephone, they do not have the capability to change the buttons or the information that is played back to them during the greeting. With this invention, users can specify a location of their greeting, therefore, depending on the user role, a more detailed greeting that contains additional information may be used. For example, an IT administrator might choose to put the number of critical escalation cases in their main mailbox greeting. In alternate aspects, it is to be appreciated that prompts (e.g., audio) can be customized on a per user basis. As well, the ordering of the state machine can user customizable.
- In summary, aspects of the subject invention are directed toward at least the following novel components employed in connection with the
UM system 602. - OOF Voicemail Greeting Integration with Email Application—The aspect that the email and voicemail systems OOF status can be content matched thereby enabling a user to set the OOF status from one place. For example, the auto-reply email body can be read to the user as the telephone greeting using TTS mechanisms.
- Calendar Access—Capability to act on and interact with calendar appointments from the telephone in a novel manner. For example, a user can accept, decline, cancel, transfer to the phone of the meeting location, and initiate an email message to the meeting participants.
- Contact List Access—Capability to access personal contact list, rather than just the company directory. Ability to send, via the telephone, an email and sound file attachment to an email account.
- PIN Reset—Allowing the user to reset a password (e.g., voicemail password) via an email client. It will be understood that the operating system security credentials can maintain security.
- Digital Rights Management for Voicemail—Digital rights for voicemails are a new and novel concept. In other words, the invention can be employed to control and/or restrict access to voicemail content. For instance, a recipient can listen to a voicemail, but can be restricted from forwarding or saving the message.
- Text Preview of Voicemails—Using speech to text conversion techniques, the subject invention can convert recorded voicemails into a textual message. The message can then be sent to the recipient via an email client application. As well, the system contemplates text-to-speech conversion for text messages. Moreover, the system contemplates generating and attaching a sound file (e.g., .wav format) of a voice message to an email. Thus, the email and attachment can be forwarded to the recipient.
- Secure Facsimile Handling—The subject invention can extract the cover page (e.g., first page) from a received facsimile transmission thereby, retaining confidentiality of remaining pages of the transmission.
- Most Importance Sentences—Based upon predefined logic and/or reasoning techniques, the invention can determine and rank sentences of a communication (e.g., email and/or voicemail) in terms of importance. For example, ranking can occur with regard to a given an email thread and key words contained therein. It will be appreciated that the ranking of sentences and/or keywords can generally provide a summary of the entire email without having to review it completely. Additionally, aspects are directed to using this technology for reading textual emails via the telephone, which is also a novel concept of the invention.
- Important Messages—The fact that users can optionally apply filtering criteria (e.g., rule-based, AI) is particularly novel with respect to email messages. The subject invention can analyze email threads and rank them accordingly. By way of example, the system can determine the sender of a message, employ a policy, logic and/or reasoning mechanism thereby ranking and/or categorizing the importance of the messages. In one aspect, the system can use this technology to determine an order with respect to reading the messages to the user over the telephone. As well, this technology can be employed to filter emails prior to presenting them to a user via a telephone.
- Missed Call Notification—This component can be integrated into the unified messaging system thereby providing third party call control by monitoring traffic. In one aspect, the system can monitor and advise a client of received calls.
- Policy and Compliance Integration with Email System—novel rule-based and/or AI policies can be applied to voicemails and/or emails thereby effecting management and retention of the content.
- Button Profiles—The subject invention provides for functionality that integrates novel button profile definition mechanisms.
- Spoken Password Voicemail Authentication—The invention can employ a secure spoken password to access a voicemail and/or email mailbox through a telephone. This audible authentication is a novel use of the authentication technology.
- End-User Configuration of Menus and Prompts—The subject invention provides for the ability to customize end-user menus and prompts.
-
FIG. 9 illustrates an exemplary flow chart of acts to employ an OOF voice message to set an email OOF notification. While, for purposes of simplicity of explanation, the one or more methodologies shown herein, e.g., in the form of a flow chart or diagram, are shown and described as a series of acts, it is to be understood and appreciated that the subject invention is not limited by the order of acts, as some acts may, in accordance with the subject invention, occur in a different order and/or concurrently with other acts from that shown and described herein. For example, those skilled in the art will understand and appreciate that a methodology could alternatively be represented as a series of interrelated states or events, such as in a state diagram. Moreover, not all illustrated acts may be required to implement a methodology in accordance with the subject invention. - At 902, an audible OOF message is recorded. The system can identify corresponding email account(s) at 904. Once identified, conversion can begin. At 906, the audible OOF message can be converted to a text OOF message. Accordingly, at 908, the text OOF message can be applied to the identified email account(s). Thus, although different format (e.g., voicemail vs. email), the content of the OOF can be replicated to the disparate system. Although the methodology of
FIG. 9 illustrates a process that can convert a voicemail OOF into an email OOF, it is to be understood that an alternate novel aspect of the subject invention can be employed to convert an email OOF into a voicemail system. - As described supra, it is to be appreciated that the aforementioned novel aspects of the invention can employ rule-based logic and/or AI reasoning technology to effect, predict and/or infer an action. More particularly, it will be appreciated that the subject invention (e.g., in connection with content analysis, content retention policy, synchronization, voice recognition) can employ various rule-based and/or AI-based schemes for carrying out various aspects thereof.
- Continuing with the above described scenarios and with reference to
FIG. 10 , a block schematic view of an alternativeUM application component 720 is shown. As illustrated,UM application component 720 can include a rule-basedlogic engine 1002. As will later be described with reference toFIG. 11 , an optional AI component (not shown) can be used together with, or in place of, the rule-basedlogic engine 1002 to automatically infer an action or set of actions to be employed in connection with the functionality of the UM application component 720 (and UM system 602) described supra. - In the exemplary aspect of
FIG. 10 , the rule-basedlogic engine 1002 can be programmed or configured in accordance with a predefined preference (e.g., rule). For example, a rule (e.g., logic) can be constructed to automatically prioritize emails based upon a defined hierarchy. In accordance therewith, the emails can be converted and read to a user in a prioritized order. More particularly, a rule can be established to take into consideration the subject, sender identification, recipient identification, etc., to prioritize and process emails. In another aspect, a rule can be established to predetermine or define a policy whereby emails and/or voicemails can be managed (e.g., retained, forwarded, deleted) in accordance with the policy. - A schematic diagram of another alternative aspect of the subject invention is illustrated in
FIG. 11 . TheUM application component 720 of this alternative aspect illustrated inFIG. 11 employs anoptional AI component 1102 that can automatically infer and/or predict an action. This alternative aspect can optionally include an inference module (not shown) that facilitates automatic control and/or operation of theUM application component 720. - In accordance with this aspect, the
optional AI component 1102 can facilitate automatically performing various aspects (e.g., analysis and prioritization of content, content retention policy, synchronization, voice recognition) of the subject invention as described herein. TheAI component 1102 can optionally include an inference component (not shown) that can further enhance automated aspects of the AI component utilizing, in part, inference based schemes to facilitate inferring intended actions to be performed at a given time and/or state. The AI-based aspects of the invention can be effected via any suitable machine-learning based technique and/or statistical-based techniques and/or probabilistic-based techniques. - In the alternate aspect, as further illustrated by
FIG. 11 , the subject invention (e.g., in connection with content analysis, content retention policy, synchronization, voice recognition) can optionally employ various artificial intelligence based schemes for automatically carrying out various aspects thereof. Specifically, theAI component 1102 can optionally be provided to implement aspects of the subject invention based upon AI processes (e.g., confidence, inference). For example, a process for determining the prioritization of content in an email based upon user preferences or sender identification can be facilitated via an automatic classifier system and process. Further, theoptional AI component 1102 can be employed to facilitate an automated process of rendering prioritized content to a user based upon a predicted and/or inferred preference. - A classifier is a function that maps an input attribute vector, x=(x1, x2, x3, x4, xn), to a confidence that the input belongs to a class, that is, f(x)=confidence(class). Such classification can employ a probabilistic and/or statistical-based analysis (e.g., factoring into the analysis utilities and costs) to prognose or infer an action that a user desires to be automatically performed.
- A support vector machine (SVM) is an example of a classifier that can be employed. The SVM operates by finding a hypersurface in the space of possible inputs, which hypersurface attempts to split the triggering criteria from the non-triggering events. Intuitively, this makes the classification correct for testing data that is near, but not identical to training data. Other directed and undirected model classification approaches include, e.g., naive Bayes, Bayesian networks, decision trees, and probabilistic classification models providing different patterns of independence can be employed. Classification as used herein also is inclusive of statistical regression that is utilized to develop models of priority.
- As will be readily appreciated from the subject specification, the subject invention can employ classifiers that are explicitly trained (e.g., via a generic training data) as well as implicitly trained (e.g., via observing user behavior, receiving extrinsic information). For example, SVM's can be configured via a learning or training phase within a classifier constructor and feature selection module. In other words, the use of expert systems, fuzzy logic, support vector machines, greedy search algorithms, rule-based systems, Bayesian models (e.g., Bayesian networks), neural networks, other non-linear training techniques, data fusion, utility-based analytical systems, systems employing Bayesian models, etc., are contemplated and are intended to fall within the scope of the hereto appended claims.
- Illustrated in
FIGS. 12 to 16 are flow diagrams of an exemplary user interface (UI) of a UM system in accordance with an aspect of the invention. As described supra, this UM system can facilitate management of voicemail, email and faxes stored in a server mailbox. The subject invention facilitates access to and management of this information from a telephone UI. In accordance with the UI, users are able to access their voicemail and other useful information from the telephone in lieu of accessing a desktop computer or handheld device. - In accordance with the invention,
FIGS. 12 to 16 are illustrative examples of a DTMF process flow in accordance with a disclosed aspect. Similarly,FIGS. 17 to 21 are illustrative examples of a speech activated process flow in accordance with a disclosed aspect. As described supra, it is to be appreciated that the DTMF and speech recognition interfaces can be active concurrently with respect to UM system navigational mechanisms. - Following are exemplary scenarios that this invention can address. In the first scenario, suppose a user is headed to a meeting in a remote location, but is not sure which room in the building the meeting will be held. The user may want to quickly find this information while driving his car toward the meeting. One aspect of this invention can facilitate telephonic retrieval of this information.
- In another scenario, suppose a user is actually late to a meeting and desires to call into the conference room to participate over the phone before getting there. Here, the user can telephonically access the meeting location information and subsequently be transferred to the meeting location.
- In a third scenario, suppose a user is driving to work in the morning and wants to get a head start on the some of the issues that are pending in emails or voicemails. From the telephone UI, the user can quickly review the different issues, reply as desired, and follow-up with the people involved with the issues.
- In a final exemplary scenario, suppose a user is in a hotel room without access to a printer and wishes to send emails as well as calendar appointments for the following day to the hotel fax so that the user can review materials and plan the day. One aspect of the invention facilitates this functionality via a telephonic connection. The procedure flow of these and other aspects will be better understood upon a review of
FIGS. 12 to 16 that follow. - Referring now to
FIG. 12 , an exemplary process task flow diagram of calling into a UM is shown. The process illustrates a flow of procedures in accordance to a call initiated by anexternal caller 1202 and aninternal caller 1204. As illustrated, the process flow of the exemplary aspect is dependent upon the type of caller (e.g., external 1202 or internal 1204) as well as the contact number dialed at 1206. An external caller will be prompted with the company information at 1208 whereby an internal caller with internal information at 1210. - In each scenario, a caller can contact a user located on the system by spelling the name and employing the directory search at 1212 or by connecting directly to the extension (1214) by entering the numeric extension. If the user does not answer, at 1216, the system can play the appropriate greeting (e.g., external, internal, OOF) in accordance with the type of caller. Accordingly, a voicemail message can be recorded at 1218. If a pre-designated key (e.g., “*”) is pressed, the system can enter the callee mailbox at 1220.
- It is to be appreciated that the callee can access the mailbox directly from 1220 by choosing a pre-designated key (e.g., “#”) from the company and/or internal greeting menu (1208, 1210). This direct access will prompt the callee for necessary information at 1222. In each case, once the appropriate security credential information is entered at 1220, access to the UM is gained at 1224.
FIG. 13 illustrates a continued procedure flow diagram once access is granted. - Referring now to
FIG. 13 , once access to the UM is granted at 1224, the user can be prompted with a main menu at 1302. By way of example, the user can be prompted to press “1” for voicemail, “2” for calendar options, “3” for contacts, “4” for email and so on. Option numbers “2” and “3”, e.g., calendar options and contacts, will be discussed in greater detail with reference toFIGS. 14 and 15 respectfully. By way of further example, a user can choose “5” to send a message (e.g., voicemail, email) or “7” to set an OOF. Additional exemplary menu choices are illustrated inblock 1302. - As shown, if voicemail or email is chosen the system proceeds to 1304 whereby content can be presented to the user. In other words, the user can be presented with voicemail, email and/or meeting request content. It is to be appreciated that logic and reasoning mechanisms can be employed to manage, organize, filter and/or sort the content as described supra. As illustrated, once the content is presented, the user can be presented with an “after message” menu thereby presenting additional options. By way of example, by pressing “4” a user can send a facsimile of a desired message. In other words, once a user listens to a voicemail and/or TTS converted email, the user can opt to send a facsimile of the message to a desired location. As well, by pressing “3”, the user can reply to the message or accept a meeting appointment request. By choosing option “2”, a user can call the sender of a particular message at 1306. This “call” option will be discussed in greater detail with reference to
FIG. 15 . It is to be appreciated that the options shown in the UI process flow ofFIG. 13 are exemplary and are not intended to be an exhaustive list of options available to a user. - At 1308, a list of exemplary play commands, which are presented during playback of a message, are shown. Again, this list is not intended to be exhaustive in view of the functionality described herein. As illustrated, it will be appreciated that the system can effect a pause, undelete, repeat, etc., task in accordance with options shown at 1306.
- Turning now to
FIG. 14 , an exemplary flow diagram of calendar access is shown. At 1402, the system can play the number of pending meetings. As well, as illustrated at 1402, the user can manage the meeting appointments and/or requests by selecting any of various options. For example, the user can select option “2” to call the meeting location, “2,2” to call the meeting organizer or “3” to indicate a late arrival to the subject meeting. As well, additional navigational options can be presented to the user as illustrated. -
FIG. 15 illustrates an exemplary process flow to effect generating a call from the contacts menu. It is to be appreciated that a call can also be generated in accordance with a meeting request/appointment or a received communication. Referring again toFIG. 15 , once a contact is located via the directory search at 1212, the system can present user information at 1502. The information presented at 1502 can effect any desired task. By way of example, the user can call or send a message to the contact by pressing “2” or “3” respectively. If the user chooses to call, the process continues to generate the call as illustrated. - Illustrated in
FIG. 16 is a process flow diagram to record a message. The message is recorded at 1602. At 1604, the user is presented with a variety of options to manage the recording. For example, the user can mark the message with high priority by depressing “1,1” or restrict access by depressing “5”. Once complete, the user can exit the system at 1606. -
FIGS. 17 to 21 illustrate exemplary process flow diagrams of a UM system that employs a speech UI. As described supra, a user can selectively employ disparate navigational mechanisms with respect to the application state machine (e.g., UM system). The process flow diagrams ofFIGS. 17 to 21 illustrate the co-existence of the DTMF UI (FIGS. 12 to 16 ) and a speech activated UI. It is to be understood that a user can employ any available UI at any time in connection with navigating within an application state machine. - Referring now to
FIG. 17 , an exemplary process task flow diagram of calling into a UM system is shown. As discussed with reference withFIG. 12 , the process illustrates a flow of procedures in accordance to a call initiated by anexternal caller 1202 and aninternal caller 1204. As illustrated, the process flow of the exemplary aspect is dependent upon the type of caller (e.g., external 1202 or internal 1204) as well as the contact number dialed at 1206. An external caller will be prompted with the company information at 1208 whereby an internal caller with internal information at 1210. - In each scenario, a caller can contact a user located on the system by employing the directory search (1212 of
FIG. 12 ) or by connecting directly to the extension (1214) by entering the numeric extension. As illustrated inFIG. 17 , the user can speak commands at 1702 to effect a desired option. As well, at 1704, the user can verbally instruct the system of a desired target person to contact. At any time, a user can speak the word “help” (1706) to access available system options.FIG. 18 illustrates a continued procedure flow diagram that includes voice options once access is granted. - Referring now to
FIG. 18 , once access to the UM system is granted at 1224, the user can be prompted with a main menu at 1302. By way of example, the user can be prompted to press “1” for voicemail, “2” for calendar options, “3” for contacts, “4” for email and so on. By way of further example, a user can choose “5” to send a message (e.g., voicemail, email) or “7” to set an OOF. Additional exemplary DTMF menu choices are illustrated inblock 1302. As well, if desired, voice commands can be employed at 1802 in order to effect navigation. - As shown, if voicemail or email is chosen the system proceeds to 1304 whereby content can be presented to the user. In other words, the user can be presented with voicemail, email and/or meeting request content. It is to be appreciated that logic and reasoning mechanisms can be employed to manage, organize, filter and/or sort the content as described supra. As illustrated, once the content is presented, the user can be presented with an “after message” menu thereby presenting additional options. By way of example, by pressing “4” a user can send a facsimile of a desired message. In other words, once a user listens to a voicemail and/or TTS converted email, the user can opt to send a facsimile of the message to a desired location. As well, by pressing “3”, the user can reply to the message or accept a meeting appointment request. By choosing option “2”, a user can call the sender of a particular message at 1306. It is to be appreciated that the options shown in the UI process flow of
FIG. 18 are exemplary and are not intended to be an exhaustive list of options available to a user. - It will further be appreciated that the voice commands shown at 1804 can be employed “during” or “after message” playback. In other words, the voice commands can be reused in either scenario. These reusable commands are listed on
FIG. 18 as “anytime” commands. - At 1308, a list of exemplary DTMF play commands, which are presented during playback of a message, are shown. Again, this list is not intended to be exhaustive in view of the functionality described herein. As illustrated, it will be appreciated that the system can effect a pause, undelete, repeat, etc., task in accordance with options shown at 1308. In one aspect at 1806, speech commands and options can be employed in connection with pausing a playback.
- Turning now to
FIG. 19 , an exemplary flow diagram of calendar access is shown. At 1402, the system can play the number of pending meetings. As well, as illustrated at 1402, the user can manage the meeting appointments and/or requests by selecting any of various options. For example, the user can select option “2” to call the meeting location, “2,2” to call the meeting organizer or “3” to indicate a late arrival to the subject meeting. As well, additional navigational options can be presented to the user as illustrated. For example, as illustrated, a user can employ speech commands at 1902 to effect calendar playback. -
FIG. 20 illustrates an exemplary process flow to effect generating a call from the contacts menu. It is to be appreciated that a call can also be generated in accordance with a meeting request/appointment or a received communication. Referring again toFIG. 20 , once a contact is verbally located via the directory search at 2002, the system can present user information at 1502. The information presented at 1502 can effect any desired task. By way of example, the user can call or send a message to the contact by pressing “2” or “3” respectively. If the user chooses to call, the process continues to generate the call as illustrated. Additionally, at 2004, the user can effect communication with a contact by employing speech commands. For example, the user can audibly instruct the system to “find” a particular person. Next, the user can audibly tell the system to send an email message to the located contact. - Illustrated in
FIG. 21 is a process flow diagram to record a message. The message is recorded at 1602. At 1604, the user is presented with a variety of options to manage the recording. For example, the user can mark the message with high priority by depressing “1,1” or restrict access by depressing “5”. Once complete, the user can exit the system at 1606. As discussed with respect to the other exemplary flow diagrams, a user at any time can opt to communicate with the system via another available UI (e.g., speech recognition). In the example ofFIG. 21 , the user can speak the commands identified at 2102 thus navigating through the application state machine (e.g., UM system). - Referring now to
FIG. 22 , there is illustrated a block diagram of a computer operable to execute the disclosed architecture. In order to provide additional context for various aspects of the subject invention,FIG. 22 and the following discussion are intended to provide a brief, general description of asuitable computing environment 2200 in which the various aspects of the subject invention can be implemented. While the invention has been described above in the general context of computer-executable instructions that may run on one or more computers, those skilled in the art will recognize that the invention also can be implemented in combination with other program modules and/or as a combination of hardware and software. - Generally, program modules include routines, programs, components, data structures, etc., that perform particular tasks or implement particular abstract data types. Moreover, those skilled in the art will appreciate that the inventive methods can be practiced with other computer system configurations, including single-processor or multiprocessor computer systems, minicomputers, mainframe computers, as well as personal computers, hand-held computing devices, microprocessor-based or programmable consumer electronics, and the like, each of which can be operatively coupled to one or more associated devices.
- The illustrated aspects of the invention may also be practiced in distributed computing environments where certain tasks are performed by remote processing devices that are linked through a communications network. In a distributed computing environment, program modules can be located in both local and remote memory storage devices.
- A computer typically includes a variety of computer-readable media. Computer-readable media can be any available media that can be accessed by the computer and includes both volatile and nonvolatile media, removable and non-removable media. By way of example, and not limitation, computer readable media can comprise computer storage media and communication media. Computer storage media includes both volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer readable instructions, data structures, program modules or other data. Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital video disk (DVD) or other optical disk storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed by the computer.
- Communication media typically embodies computer-readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism, and includes any information delivery media. The term “modulated data signal” means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal. By way of example, and not limitation, communication media includes wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared and other wireless media. Combinations of the any of the above should also be included within the scope of computer-readable media.
- With reference again to
FIG. 22 , there is illustrated anexemplary environment 2200 for implementing various aspects of the invention that includes acomputer 2202, thecomputer 2202 including aprocessing unit 2204, asystem memory 2206 and asystem bus 2208. Thesystem bus 2208 couples system components including, but not limited to, thesystem memory 2206 to theprocessing unit 2204. Theprocessing unit 2204 can be any of various commercially available processors. Dual microprocessors and other multi-processor architectures may also be employed as theprocessing unit 2204. - The
system bus 2208 can be any of several types of bus structure that may further interconnect to a memory bus (with or without a memory controller), a peripheral bus, and a local bus using any of a variety of commercially available bus architectures. Thesystem memory 2206 includes read only memory (ROM) 2210 and random access memory (RAM) 2212. A basic input/output system (BIOS) is stored in anon-volatile memory 2210 such as ROM, EPROM, EEPROM, which BIOS contains the basic routines that help to transfer information between elements within thecomputer 2202, such as during start-up. TheRAM 2212 can also include a high-speed RAM such as static RAM for caching data. - The
computer 2202 further includes an internal hard disk drive (HDD) 2214 (e.g., EIDE, SATA), which internalhard disk drive 2214 may also be configured for external use in a suitable chassis (not shown), a magnetic floppy disk drive (FDD) 2216, (e.g., to read from or write to a removable diskette 2218) and anoptical disk drive 2220, (e.g., reading a CD-ROM disk 2222 or, to read from or write to other high capacity optical media such as the DVD). Thehard disk drive 2214,magnetic disk drive 2216 andoptical disk drive 2220 can be connected to thesystem bus 2208 by a harddisk drive interface 2224, a magneticdisk drive interface 2226 and anoptical drive interface 2228, respectively. Theinterface 2224 for external drive implementations includes at least one or both of Universal Serial Bus (USB) and IEEE 1394 interface technologies. - The drives and their associated computer-readable media provide nonvolatile storage of data, data structures, computer-executable instructions, and so forth. For the
computer 2202, the drives and media accommodate the storage of any data in a suitable digital format. Although the description of computer-readable media above refers to a HDD, a removable magnetic diskette, and a removable optical media such as a CD or DVD, it should be appreciated by those skilled in the art that other types of media which are readable by a computer, such as zip drives, magnetic cassettes, flash memory cards, cartridges, and the like, may also be used in the exemplary operating environment, and further, that any such media may contain computer-executable instructions for performing the methods of the subject invention. - A number of program modules can be stored in the drives and
RAM 2212, including anoperating system 2230, one or more application programs 2232,other program modules 2234 andprogram data 2236. All or portions of the operating system, applications, modules, and/or data can also be cached in theRAM 2212. It is appreciated that the subject invention can be implemented with various commercially available operating systems or combinations of operating systems. - A user can enter commands and information into the
computer 2202 through one or more wired/wireless input devices, e.g., akeyboard 2238 and a pointing device, such as amouse 2240. Other input devices (not shown) may include a microphone, an IR remote control, a joystick, a game pad, a stylus pen, touch screen, or the like. These and other input devices are often connected to theprocessing unit 2204 through aninput device interface 2242 that is coupled to thesystem bus 2208, but can be connected by other interfaces, such as a parallel port, an IEEE 1394 serial port, a game port, a USB port, an IR interface, etc. - A
monitor 2244 or other type of display device is also connected to thesystem bus 2208 via an interface, such as avideo adapter 2246. In addition to themonitor 2244, a computer typically includes other peripheral output devices (not shown), such as speakers, printers, etc. - The
computer 2202 may operate in a networked environment using logical connections via wired and/or wireless communications to one or more remote computers, such as a remote computer(s) 2248. The remote computer(s) 2248 can be a workstation, a server computer, a router, a personal computer, portable computer, microprocessor-based entertainment appliance, a peer device or other common network node, and typically includes many or all of the elements described relative to thecomputer 2202, although, for purposes of brevity, only amemory storage device 2250 is illustrated. The logical connections depicted include wired/wireless connectivity to a local area network (LAN) 2252 and/or larger networks, e.g., a wide area network (WAN) 2254. Such LAN and WAN networking environments are commonplace in offices, and companies, and facilitate enterprise-wide computer networks, such as intranets, all of which may connect to a global communication network, e.g., the Internet. - When used in a LAN networking environment, the
computer 2202 is connected to the local network 2252 through a wired and/or wireless communication network interface oradapter 2256. Theadaptor 2256 may facilitate wired or wireless communication to the LAN 2252, which may also include a wireless access point disposed thereon for communicating with thewireless adaptor 2256. When used in a WAN networking environment, thecomputer 2202 can include amodem 2258, or is connected to a communications server on theWAN 2254, or has other means for establishing communications over theWAN 2254, such as by way of the Internet. Themodem 2258, which can be internal or external and a wired or wireless device, is connected to thesystem bus 2208 via theserial port interface 2242. In a networked environment, program modules depicted relative to thecomputer 2202, or portions thereof, can be stored in the remote memory/storage device 2250. It will be appreciated that the network connections shown are exemplary and other means of establishing a communications link between the computers can be used. - The
computer 2202 is operable to communicate with any wireless devices or entities operatively disposed in wireless communication, e.g., a printer, scanner, desktop and/or portable computer, portable data assistant, communications satellite, any piece of equipment or location associated with a wirelessly detectable tag (e.g., a kiosk, news stand, restroom), and telephone. This includes at least Wi-Fi and Bluetooth™. wireless technologies. Thus, the communication can be a predefined structure as with conventional network or simply an ad hoc communication between at least two devices. - Wi-Fi, or Wireless Fidelity, allows connection to the Internet from a couch at home, a bed in a hotel room or a conference room at work, without wires. Wi-Fi is a wireless technology like a cell phone that enables such devices, e.g., computers, to send and receive data indoors and out; anywhere within the range of a base station. Wi-Fi networks use radio technologies called IEEE 802.11 (a, b, g, etc.) to provide secure, reliable, fast wireless connectivity. A Wi-Fi network can be used to connect computers to each other, to the Internet, and to wired networks (which use IEEE 802.3 or Ethernet). Wi-Fi networks operate in the unlicensed 2.4 and 5 GHz radio bands, at an 11 Mbps (802.11 a) or 54 Mbps (802.11 b) data rate, for example, or with products that contain both bands (dual band), so the networks can provide real-world performance similar to the basic 10 BaseT wired Ethernet networks used in many offices.
- Referring now to
FIG. 23 , there is illustrated a schematic block diagram of anexemplary computing environment 2300 in accordance with the subject invention. Thesystem 2300 includes one or more client(s) 2302. The client(s) 2302 can be hardware and/or software (e.g., threads, processes, computing devices). The client(s) 2302 can house cookie(s) and/or associated contextual information by employing the subject invention, for example. Thesystem 2300 also includes one or more server(s) 2304. The server(s) 2304 can also be hardware and/or software (e.g., threads, processes, computing devices). Theservers 2304 can house threads to perform transformations by employing the subject invention, for example. One possible communication between aclient 2302 and aserver 2304 can be in the form of a data packet adapted to be transmitted between two or more computer processes. The data packet may include a cookie and/or associated contextual information, for example. Thesystem 2300 includes a communication framework 2306 (e.g., a global communication network such as the Internet) that can be employed to facilitate communications between the client(s) 2302 and the server(s) 2304. - Communications can be facilitated via a wired (including optical fiber) and/or wireless technology. The client(s) 2302 are operatively connected to one or more client data store(s) 2308 that can be employed to store information local to the client(s) 2302 (e.g., cookie(s) and/or associated contextual information). Similarly, the server(s) 2304 are operatively connected to one or more server data store(s) 2310 that can be employed to store information local to the
servers 2304. - What has been described above includes examples of the subject invention. It is, of course, not possible to describe every conceivable combination of components or methodologies for purposes of describing the subject invention, but one of ordinary skill in the art may recognize that many further combinations and permutations of the subject invention are possible. Accordingly, the subject invention is intended to embrace all such alterations, modifications and variations that fall within the spirit and scope of the appended claims. Furthermore, to the extent that the term “includes” is used in either the detailed description or the claims, such term is intended to be inclusive in a manner similar to the term “comprising” as “comprising” is interpreted when employed as a transitional word in a claim.
Claims (20)
1. A method of employing a unified messaging system, the method comprising:
concurrently providing a plurality of user interfaces that facilitate navigating through the unified messaging system;
selecting one of the plurality of user interfaces;
transmitting an instruction to the unified messaging system via the selected user interface; and
processing the instruction to control operation of the unified messaging system.
2. The method of claim 1 , further comprising alternatively selecting another of the plurality of user interfaces and transmitting a second instruction in accordance with the alternative user interface.
3. The method of claim 1 , wherein processing the instruction comprises:
receiving a message of a first format from a first device;
analyzing the message;
transforming the message to a second format that corresponds to a second device; and
synchronizing the messages of the first format and the second format.
4. The method of claim 3 , wherein receiving the message of the first format comprises receiving a voicemail communication.
5. The method of claim 3 , wherein transforming the message to the second format comprises transforming the message to an email communication.
6. The method of claim 3 , further comprising conveying content of the message.
7. The method of claim 6 , wherein conveying the content of the message comprises employing at least one of the following: rule-based logic and artificial intelligence.
8. The method of claim 7 , wherein employing at least one of the following: the rule-based logic and the artificial intelligence comprises at least one of the following:
analyzing the content of the message, and
ranking the message according to importance.
9. The method of claim 1 , further comprising:
receiving audible information from an individual via telephone; and
reformatting the information.
10. The method of claim 9 , further comprising populating a calendar with the reformatted information.
11. The method of claim 1 , further comprising:
receiving a first navigation command via a first navigation input mechanism; and
selecting a first user interface of the plurality of user interfaces based on a first type of the first navigation input mechanism, the first user interface being associated with the first navigation input mechanism.
12. The method of claim 11 , further comprising entering a first state within the unified messaging system via the first user interface in accordance with the first navigation command.
13. The method of claim 12 , further comprising:
receiving a second navigation command via a second navigation input mechanism, the second navigation input mechanism being a second type different from the first type; and
selecting a second user interface of the plurality of user interfaces based on the second type of the second navigation input mechanism, the second user interface being associated with the second navigation input mechanism.
14. The method of claim 13 , further comprising entering a second state within the unified messaging system via the second user interface in accordance with the second navigation command, the second state being different from the first state.
15. The method claim 1 , further comprising conveying, by the unified messaging system, available commands associated with the selected one of the plurality of user interfaces.
16. The method of claim 15 , wherein conveying, by the unified messaging system, the available commands comprises conveying the available commands based on an input mechanism used to interact with the unified messaging system.
17. A system that facilitates navigating within an application state machine, the system comprising:
a processor;
a user interface store that retains a plurality of user interfaces that are each associated with a navigation input mechanism;
a user interface selection component, executed by the processor, that facilitates selecting one of the plurality of user; and
a user interface component, executed by the processor, that conveys a distinct set of prompts and menus associated with each user interface, wherein the distinct set of prompts and menus associated with each user interface change based on the navigation input mechanism.
18. The system of claim 17 , wherein the user interface component is configured to provide, upon first entry into the application state machine:
a default user interface based on a set personal option, and
a corresponding distinct set of prompts and menus associated with the default user interface, the corresponding distinct set of prompts being cooperative with both: a dual tone multi-frequency input and a microphone input.
19. The system of claim 18 , wherein the plurality of user interfaces is concurrently available during a single navigation session to effect navigation within the application state machine.
20. A computer-readable medium having a set of instructions which when executed performs a method for facilitates navigating within an application state machine, the method executed by the set of instructions comprising:
presenting, upon first entry into the application state machine, a default user interface mechanism of a plurality of user interface mechanisms;
enable, at any point within the application state machine, user interaction with any one of a plurality of user interface mechanisms that facilitate navigating through the application state machine, each of the plurality of user interface mechanisms being associated with a different navigation input mechanism of a plurality of navigation input mechanisms and having a distinct set of prompts and menus;
receive a navigation command via a navigation input mechanism of the plurality of navigation input mechanisms;
select a user interface mechanism of the plurality of user interface mechanisms based on a type of the navigation input mechanism;
receive an instruction via the user interface mechanism; and
process the instruction to control the application state machine.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/044,581 US20110216889A1 (en) | 2004-10-20 | 2011-03-10 | Selectable State Machine User Interface System |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US62082604P | 2004-10-20 | 2004-10-20 | |
US11/086,828 US7912186B2 (en) | 2004-10-20 | 2005-03-22 | Selectable state machine user interface system |
US13/044,581 US20110216889A1 (en) | 2004-10-20 | 2011-03-10 | Selectable State Machine User Interface System |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/086,828 Continuation US7912186B2 (en) | 2004-10-20 | 2005-03-22 | Selectable state machine user interface system |
Publications (1)
Publication Number | Publication Date |
---|---|
US20110216889A1 true US20110216889A1 (en) | 2011-09-08 |
Family
ID=36499232
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/086,828 Expired - Fee Related US7912186B2 (en) | 2004-10-20 | 2005-03-22 | Selectable state machine user interface system |
US13/044,581 Abandoned US20110216889A1 (en) | 2004-10-20 | 2011-03-10 | Selectable State Machine User Interface System |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/086,828 Expired - Fee Related US7912186B2 (en) | 2004-10-20 | 2005-03-22 | Selectable state machine user interface system |
Country Status (5)
Country | Link |
---|---|
US (2) | US7912186B2 (en) |
EP (1) | EP1705886A1 (en) |
JP (1) | JP2006268832A (en) |
KR (1) | KR20060102268A (en) |
CN (1) | CN1838694A (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8090083B2 (en) | 2004-10-20 | 2012-01-03 | Microsoft Corporation | Unified messaging architecture |
US20120194877A1 (en) * | 2003-02-11 | 2012-08-02 | Omtool, Inc. | Method and system for secure facsimile delivery and registration |
US20120281819A1 (en) * | 2011-05-05 | 2012-11-08 | Shreedhar Madhavapeddi | Voicemail preview and editing system |
US8904270B2 (en) | 2006-11-29 | 2014-12-02 | Omtool Ltd. | Methods and apparatus for enterprise document distribution |
US11212384B2 (en) | 2019-12-20 | 2021-12-28 | Capital One Services, Llc | Cross-channel configurable navigation |
Families Citing this family (67)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6985943B2 (en) * | 1998-09-11 | 2006-01-10 | Genesys Telecommunications Laboratories, Inc. | Method and apparatus for extended management of state and interaction of a remote knowledge worker from a contact center |
US7580837B2 (en) | 2004-08-12 | 2009-08-25 | At&T Intellectual Property I, L.P. | System and method for targeted tuning module of a speech recognition system |
US7912186B2 (en) * | 2004-10-20 | 2011-03-22 | Microsoft Corporation | Selectable state machine user interface system |
US7242751B2 (en) | 2004-12-06 | 2007-07-10 | Sbc Knowledge Ventures, L.P. | System and method for speech recognition-enabled automatic call routing |
US7864942B2 (en) * | 2004-12-06 | 2011-01-04 | At&T Intellectual Property I, L.P. | System and method for routing calls |
US7751551B2 (en) | 2005-01-10 | 2010-07-06 | At&T Intellectual Property I, L.P. | System and method for speech-enabled call routing |
US7657020B2 (en) | 2005-06-03 | 2010-02-02 | At&T Intellectual Property I, Lp | Call routing system and method of using the same |
US7689253B2 (en) | 2005-06-13 | 2010-03-30 | E-Lane Systems, Inc. | Vehicle immersive communication system |
US7773731B2 (en) * | 2005-12-14 | 2010-08-10 | At&T Intellectual Property I, L. P. | Methods, systems, and products for dynamically-changing IVR architectures |
US7577664B2 (en) | 2005-12-16 | 2009-08-18 | At&T Intellectual Property I, L.P. | Methods, systems, and products for searching interactive menu prompting system architectures |
US9008075B2 (en) | 2005-12-22 | 2015-04-14 | Genesys Telecommunications Laboratories, Inc. | System and methods for improving interaction routing performance |
US8315874B2 (en) * | 2005-12-30 | 2012-11-20 | Microsoft Corporation | Voice user interface authoring tool |
US9275129B2 (en) | 2006-01-23 | 2016-03-01 | Symantec Corporation | Methods and systems to efficiently find similar and near-duplicate emails and files |
US8392409B1 (en) | 2006-01-23 | 2013-03-05 | Symantec Corporation | Methods, systems, and user interface for E-mail analysis and review |
US7657603B1 (en) * | 2006-01-23 | 2010-02-02 | Clearwell Systems, Inc. | Methods and systems of electronic message derivation |
US7899871B1 (en) | 2006-01-23 | 2011-03-01 | Clearwell Systems, Inc. | Methods and systems for e-mail topic classification |
US9600568B2 (en) | 2006-01-23 | 2017-03-21 | Veritas Technologies Llc | Methods and systems for automatic evaluation of electronic discovery review and productions |
US7743051B1 (en) | 2006-01-23 | 2010-06-22 | Clearwell Systems, Inc. | Methods, systems, and user interface for e-mail search and retrieval |
US8364467B1 (en) * | 2006-03-31 | 2013-01-29 | Google Inc. | Content-based classification |
US20070274468A1 (en) * | 2006-05-11 | 2007-11-29 | Lucent Technologies, Inc. | Retrieval of voicemail |
MX2009000006A (en) * | 2006-06-30 | 2009-04-06 | Vonage Network Llc | Method and apparatus for visual message indication in a voip system. |
US8495147B1 (en) * | 2006-07-13 | 2013-07-23 | Avaya Inc. | Threading of mixed media |
US9976865B2 (en) | 2006-07-28 | 2018-05-22 | Ridetones, Inc. | Vehicle communication system with navigation |
US20080095333A1 (en) * | 2006-10-18 | 2008-04-24 | Voicebutton, Llc | System and method of communicating internet user information to service providers |
US7948984B2 (en) * | 2007-01-16 | 2011-05-24 | Cisco Technology, Inc. | Detecting an identifier change of an endpoint |
US9092434B2 (en) | 2007-01-23 | 2015-07-28 | Symantec Corporation | Systems and methods for tagging emails by discussions |
US8695074B2 (en) * | 2007-04-26 | 2014-04-08 | Microsoft Corporation | Pre-authenticated calling for voice applications |
US8549412B2 (en) | 2007-07-25 | 2013-10-01 | Yahoo! Inc. | Method and system for display of information in a communication system gathered from external sources |
US8074199B2 (en) * | 2007-09-24 | 2011-12-06 | Microsoft Corporation | Unified messaging state machine |
US8359204B2 (en) * | 2007-10-26 | 2013-01-22 | Honda Motor Co., Ltd. | Free-speech command classification for car navigation system |
US8274935B2 (en) * | 2008-02-11 | 2012-09-25 | Amdocs Software Systems Limited | Mobile communication device notifier |
CA2719301C (en) * | 2008-03-25 | 2016-10-04 | E-Lane Systems Inc. | Multi-participant, mixed-initiative voice interaction system |
TWI385932B (en) | 2008-03-26 | 2013-02-11 | Asustek Comp Inc | Device and system for remote controlling |
US8171393B2 (en) * | 2008-04-16 | 2012-05-01 | Clearwell Systems, Inc. | Method and system for producing and organizing electronically stored information |
CA2727951A1 (en) | 2008-06-19 | 2009-12-23 | E-Lane Systems Inc. | Communication system with voice mail access and call by spelling functionality |
US9652023B2 (en) | 2008-07-24 | 2017-05-16 | Intelligent Mechatronic Systems Inc. | Power management system |
US8345832B2 (en) * | 2009-01-09 | 2013-01-01 | Microsoft Corporation | Enhanced voicemail usage through automatic voicemail preview |
US8577543B2 (en) * | 2009-05-28 | 2013-11-05 | Intelligent Mechatronic Systems Inc. | Communication system with personal information management and remote vehicle monitoring and control features |
WO2010148518A1 (en) | 2009-06-27 | 2010-12-29 | Intelligent Mechatronic Systems | Vehicle internet radio interface |
US9721228B2 (en) | 2009-07-08 | 2017-08-01 | Yahoo! Inc. | Locally hosting a social network using social data stored on a user's computer |
US8935656B2 (en) * | 2009-09-09 | 2015-01-13 | International Business Machines Corporation | Communicating information in computing systems |
US9514466B2 (en) | 2009-11-16 | 2016-12-06 | Yahoo! Inc. | Collecting and presenting data including links from communications sent to or from a user |
US9978272B2 (en) | 2009-11-25 | 2018-05-22 | Ridetones, Inc | Vehicle to vehicle chatting and communication system |
US9760866B2 (en) | 2009-12-15 | 2017-09-12 | Yahoo Holdings, Inc. | Systems and methods to provide server side profile information |
US20110153668A1 (en) * | 2009-12-18 | 2011-06-23 | Research In Motion Limited | Accessing a data item stored in an unavailable mobile communication device |
EP2337325B1 (en) * | 2009-12-18 | 2014-07-30 | BlackBerry Limited | Accessing a data item stored in an unavailable mobile communication device |
US20110211679A1 (en) * | 2010-02-26 | 2011-09-01 | Vladimir Mezhibovsky | Voice Response Processing |
US8972257B2 (en) | 2010-06-02 | 2015-03-03 | Yahoo! Inc. | Systems and methods to present voice message information to a user of a computing device |
US8620935B2 (en) | 2011-06-24 | 2013-12-31 | Yahoo! Inc. | Personalizing an online service based on data collected for a user of a computing device |
US8719257B2 (en) | 2011-02-16 | 2014-05-06 | Symantec Corporation | Methods and systems for automatically generating semantic/concept searches |
US9491291B1 (en) * | 2011-04-20 | 2016-11-08 | Confinement Telephony Technology, Llc | Systems and methods for institutional messaging |
US10078819B2 (en) | 2011-06-21 | 2018-09-18 | Oath Inc. | Presenting favorite contacts information to a user of a computing device |
US9747583B2 (en) | 2011-06-30 | 2017-08-29 | Yahoo Holdings, Inc. | Presenting entity profile information to a user of a computing device |
US9143910B2 (en) | 2011-09-30 | 2015-09-22 | Blackberry Limited | Method and system for remote wipe through voice mail |
US10977285B2 (en) | 2012-03-28 | 2021-04-13 | Verizon Media Inc. | Using observations of a person to determine if data corresponds to the person |
US20140088971A1 (en) * | 2012-08-20 | 2014-03-27 | Michael D. Metcalf | System And Method For Voice Operated Communication Assistance |
US10013672B2 (en) | 2012-11-02 | 2018-07-03 | Oath Inc. | Address extraction from a communication |
US10192200B2 (en) | 2012-12-04 | 2019-01-29 | Oath Inc. | Classifying a portion of user contact data into local contacts |
US11237705B2 (en) * | 2013-03-15 | 2022-02-01 | Bmc Software, Inc. | Configurable user interface interactions using a state machine combined with event orchestration |
US9953646B2 (en) | 2014-09-02 | 2018-04-24 | Belleau Technologies | Method and system for dynamic speech recognition and tracking of prewritten script |
KR102340234B1 (en) * | 2014-12-23 | 2022-01-18 | 엘지전자 주식회사 | Portable device and its control method |
US9820108B1 (en) | 2015-10-20 | 2017-11-14 | Allstate Insurance Company | Connected services configurator |
US10542142B2 (en) | 2017-10-27 | 2020-01-21 | Stacey Hengsterman | Apparatus and method for accessing contact lists on an electronic device that is unavailable or unusable |
US10991255B2 (en) | 2018-04-05 | 2021-04-27 | Ge Aviation Systems Llc | Providing an open interface to a flight management system |
WO2019213372A1 (en) * | 2018-05-02 | 2019-11-07 | Cloud People Llc | Decentralized digital communication platform system and method |
US11368443B2 (en) | 2018-05-02 | 2022-06-21 | Cloud People Llc | Decentralized digital communication platform system and method |
CN113163154B (en) * | 2021-04-29 | 2022-08-23 | 北京字节跳动网络技术有限公司 | Information processing method and device and electronic equipment |
Citations (56)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5003575A (en) * | 1987-12-24 | 1991-03-26 | Chamberlin David B | Method and apparatus for storing and forwarding voice signals with controlled access |
US5557659A (en) * | 1993-06-22 | 1996-09-17 | Hyde-Thomson; Henry C. A. | Electronic mail system having integrated voice messages |
US5674002A (en) * | 1995-11-27 | 1997-10-07 | Powell; Talmadge Wayne | Hammers with optimal claw shape and method for development of optimally designed hammers, crowbars, and levers |
US5802526A (en) * | 1995-11-15 | 1998-09-01 | Microsoft Corporation | System and method for graphically displaying and navigating through an interactive voice response menu |
US5915001A (en) * | 1996-11-14 | 1999-06-22 | Vois Corporation | System and method for providing and using universally accessible voice and speech data files |
US6058415A (en) * | 1997-07-24 | 2000-05-02 | Intervoice Limited Partnership | System and method for integration of communication systems with computer-based information systems |
US6061433A (en) * | 1995-10-19 | 2000-05-09 | Intervoice Limited Partnership | Dynamically changeable menus based on externally available data |
US6064723A (en) * | 1994-09-16 | 2000-05-16 | Octel Communications Corporation | Network-based multimedia communications and directory system and method of operation |
US6072862A (en) * | 1996-07-02 | 2000-06-06 | Srinivasan; Thiru | Adaptable method and system for message delivery |
US6088428A (en) * | 1991-12-31 | 2000-07-11 | Digital Sound Corporation | Voice controlled messaging system and processing method |
US6101473A (en) * | 1997-08-08 | 2000-08-08 | Board Of Trustees, Leland Stanford Jr., University | Using speech recognition to access the internet, including access via a telephone |
US6104788A (en) * | 1997-12-04 | 2000-08-15 | Siemens Information And Communication Networks, Inc. | Apparatus and method for using a telephone for remote scheduling |
US6173043B1 (en) * | 1994-10-31 | 2001-01-09 | Worldvoice Licensing, Inc. | Telephonic voice message transmission control method |
US20010014590A1 (en) * | 2000-01-19 | 2001-08-16 | Akihiro Nakano | Portable telephone apparatus and stablly supplying method of reference frequency |
US20010015972A1 (en) * | 2000-02-21 | 2001-08-23 | Shoichi Horiguchi | Information distributing method, information distributing system, information distributing server, mobile communication network system and communication service providing method |
US20020055350A1 (en) * | 2000-07-20 | 2002-05-09 | Ash Gupte | Apparatus and method of toggling between text messages and voice messages with a wireless communication device |
US20020077829A1 (en) * | 2000-12-19 | 2002-06-20 | Brennan Paul Michael | Speech based status and control user interface customisable by the user |
US6411685B1 (en) * | 1999-01-29 | 2002-06-25 | Microsoft Corporation | System and method for providing unified messaging to a user with a thin web browser |
US20020082030A1 (en) * | 2000-11-29 | 2002-06-27 | Stefan Berndt | Apparatus and method for forwarding messages to terminals of various communication media |
US20020109731A1 (en) * | 2001-02-10 | 2002-08-15 | International Business Machines Corporation | Facilitated user interface |
US6522726B1 (en) * | 1997-03-24 | 2003-02-18 | Avaya Technology Corp. | Speech-responsive voice messaging system and method |
US20030046346A1 (en) * | 2001-07-11 | 2003-03-06 | Kirusa, Inc. | Synchronization among plural browsers |
US20030046401A1 (en) * | 2000-10-16 | 2003-03-06 | Abbott Kenneth H. | Dynamically determing appropriate computer user interfaces |
US20030059000A1 (en) * | 2001-09-27 | 2003-03-27 | Burton David Kearney | Integrated voice access to a variety of personal information services |
US6547002B1 (en) * | 2000-04-17 | 2003-04-15 | Weatherford/Lamb, Inc. | High pressure rotating drilling head assembly with hydraulically removable packer |
US20030162561A1 (en) * | 2002-02-27 | 2003-08-28 | Greg Johnson | System and method for concurrent multimodal communication session persistence |
US20030182622A1 (en) * | 2002-02-18 | 2003-09-25 | Sandeep Sibal | Technique for synchronizing visual and voice browsers to enable multi-modal browsing |
US20030195937A1 (en) * | 2002-04-16 | 2003-10-16 | Kontact Software Inc. | Intelligent message screening |
US6640242B1 (en) * | 1999-01-29 | 2003-10-28 | Microsoft Corporation | Voice access through a data-centric network to an integrated message storage and retrieval system |
US6651042B1 (en) * | 2000-06-02 | 2003-11-18 | International Business Machines Corporation | System and method for automatic voice message processing |
US20030235282A1 (en) * | 2002-02-11 | 2003-12-25 | Sichelman Ted M. | Automated transportation call-taking system |
US6697458B1 (en) * | 2000-07-10 | 2004-02-24 | Ulysses Esd, Inc. | System and method for synchronizing voice mailbox with e-mail box |
US6704394B1 (en) * | 1998-03-25 | 2004-03-09 | International Business Machines Corporation | System and method for accessing voice mail from a remote server |
US6738462B1 (en) * | 2000-07-19 | 2004-05-18 | Avaya Technology Corp. | Unified communications automated personal name addressing |
US20040141594A1 (en) * | 2003-01-20 | 2004-07-22 | Brunson Gordon R. | Messaging advise in presence-aware networks |
US20040148356A1 (en) * | 2002-11-04 | 2004-07-29 | Bishop James William | System and method for private messaging |
US20040156484A1 (en) * | 2003-02-12 | 2004-08-12 | Umesh Amin | Interface for a multi-messaging system |
US6778644B1 (en) * | 2001-12-28 | 2004-08-17 | Vocada, Inc. | Integration of voice messaging and data systems |
US6795536B1 (en) * | 1999-02-26 | 2004-09-21 | Mitel, Inc. | Automatic user preference selection for message playback based on caller line identification data |
US6801763B2 (en) * | 1997-10-29 | 2004-10-05 | Metro One Telecommunications, Inc. | Technique for effectively communicating travel directions |
US20050054381A1 (en) * | 2003-09-05 | 2005-03-10 | Samsung Electronics Co., Ltd. | Proactive user interface |
US6920425B1 (en) * | 2000-05-16 | 2005-07-19 | Nortel Networks Limited | Visual interactive response system and method translated from interactive voice response for telephone utility |
US6981023B1 (en) * | 1999-03-09 | 2005-12-27 | Michael Hamilton | Message routing |
US6987840B1 (en) * | 2001-11-06 | 2006-01-17 | At&T Corp. | Integrated message management method and system |
US20060083357A1 (en) * | 2004-10-20 | 2006-04-20 | Microsoft Corporation | Selectable state machine user interface system |
US20060083358A1 (en) * | 2004-10-20 | 2006-04-20 | Microsoft Corporation | Unified messaging architecture |
US7054939B2 (en) * | 2001-06-28 | 2006-05-30 | Bellsouth Intellectual Property Corportion | Simultaneous visual and telephonic access to interactive information delivery |
US7068643B1 (en) * | 2000-11-03 | 2006-06-27 | Intervoice Limited Partnership | Extensible interactive voice response |
US7080315B1 (en) * | 2000-06-28 | 2006-07-18 | International Business Machines Corporation | Method and apparatus for coupling a visual browser to a voice browser |
US7120234B1 (en) * | 1999-12-29 | 2006-10-10 | Bellsouth Intellectual Property Corp. | Integrated tone-based and voice-based telephone user interface |
US7136478B1 (en) * | 2004-01-13 | 2006-11-14 | Avaya Technology Corp. | Interactive voice response unit response display |
US7149777B1 (en) * | 2000-01-20 | 2006-12-12 | Avaya Technology Corp. | Method for administering message attachments |
US7283808B2 (en) * | 2001-01-18 | 2007-10-16 | Research In Motion Limited | System, method and mobile device for remote control of a voice mail system |
US7289606B2 (en) * | 2001-10-01 | 2007-10-30 | Sandeep Sibal | Mode-swapping in multi-modal telephonic applications |
US7315613B2 (en) * | 2002-03-11 | 2008-01-01 | International Business Machines Corporation | Multi-modal messaging |
US7356130B2 (en) * | 2001-10-11 | 2008-04-08 | International Business Machines Corporation | Method and system for selecting speech or DTMF interfaces or a mixture of both |
Family Cites Families (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS6041863A (en) * | 1983-08-17 | 1985-03-05 | Fujitsu Ltd | Pushphone voice input device |
US5647002A (en) | 1995-09-01 | 1997-07-08 | Lucent Technologies Inc. | Synchronization of mailboxes of different types |
WO2000018100A2 (en) | 1998-09-24 | 2000-03-30 | Crossmedia Networks Corporation | Interactive voice dialog application platform and methods for using the same |
AU1072000A (en) | 1998-11-04 | 2000-05-22 | Varicom Communications Ltd | A method for storing and forwarding voice and facsimile messages using an e-mailplatform |
WO2000054488A1 (en) | 1999-03-09 | 2000-09-14 | Michael Hamilton | Message routing |
JP2000278388A (en) * | 1999-03-29 | 2000-10-06 | Nec Saitama Ltd | Portable communication apparatus having speech recognizing function |
US7035666B2 (en) * | 1999-06-09 | 2006-04-25 | Shimon Silberfening | Combination cellular telephone, sound storage device, and email communication device |
CA2323686A1 (en) | 1999-12-23 | 2001-06-23 | Nortel Networks Limited | Methods for addressing a message from a telephone |
JP2001292240A (en) | 2000-04-06 | 2001-10-19 | Mitsubishi Electric Corp | System and device for voice mail and voice mail transfer method |
JP2002269214A (en) | 2001-03-08 | 2002-09-20 | Yoshihiko Komatsu | Schedule managing system and method using telephone set |
US7061928B2 (en) | 2001-03-26 | 2006-06-13 | Azurn Networks, Inc. | Unified XML voice and data media converging switch and application delivery system |
US7610547B2 (en) * | 2001-05-04 | 2009-10-27 | Microsoft Corporation | Markup language extensions for web enabled recognition |
DE10251112A1 (en) * | 2002-11-02 | 2004-05-19 | Philips Intellectual Property & Standards Gmbh | Voice recognition involves generating alternative recognition results during analysis with next highest probability of match to detected voice signal for use to correct output recognition result |
JP2004260416A (en) | 2003-02-25 | 2004-09-16 | Nec Access Technica Ltd | Automatic response system of cellular phone |
CN1798210B (en) | 2004-12-30 | 2010-08-11 | 国际商业机器公司 | Method and system for selecting speech or dtmf interfaces or a mixture of both |
-
2005
- 2005-03-22 US US11/086,828 patent/US7912186B2/en not_active Expired - Fee Related
-
2006
- 2006-01-31 KR KR1020060009316A patent/KR20060102268A/en not_active Application Discontinuation
- 2006-02-10 JP JP2006034135A patent/JP2006268832A/en active Pending
- 2006-02-23 CN CNA2006100093162A patent/CN1838694A/en active Pending
- 2006-03-21 EP EP06111468A patent/EP1705886A1/en not_active Withdrawn
-
2011
- 2011-03-10 US US13/044,581 patent/US20110216889A1/en not_active Abandoned
Patent Citations (60)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5003575A (en) * | 1987-12-24 | 1991-03-26 | Chamberlin David B | Method and apparatus for storing and forwarding voice signals with controlled access |
US6088428A (en) * | 1991-12-31 | 2000-07-11 | Digital Sound Corporation | Voice controlled messaging system and processing method |
US5557659A (en) * | 1993-06-22 | 1996-09-17 | Hyde-Thomson; Henry C. A. | Electronic mail system having integrated voice messages |
US6064723A (en) * | 1994-09-16 | 2000-05-16 | Octel Communications Corporation | Network-based multimedia communications and directory system and method of operation |
US6173043B1 (en) * | 1994-10-31 | 2001-01-09 | Worldvoice Licensing, Inc. | Telephonic voice message transmission control method |
US6061433A (en) * | 1995-10-19 | 2000-05-09 | Intervoice Limited Partnership | Dynamically changeable menus based on externally available data |
US5802526A (en) * | 1995-11-15 | 1998-09-01 | Microsoft Corporation | System and method for graphically displaying and navigating through an interactive voice response menu |
US5674002A (en) * | 1995-11-27 | 1997-10-07 | Powell; Talmadge Wayne | Hammers with optimal claw shape and method for development of optimally designed hammers, crowbars, and levers |
US6072862A (en) * | 1996-07-02 | 2000-06-06 | Srinivasan; Thiru | Adaptable method and system for message delivery |
US5915001A (en) * | 1996-11-14 | 1999-06-22 | Vois Corporation | System and method for providing and using universally accessible voice and speech data files |
US6522726B1 (en) * | 1997-03-24 | 2003-02-18 | Avaya Technology Corp. | Speech-responsive voice messaging system and method |
US6058415A (en) * | 1997-07-24 | 2000-05-02 | Intervoice Limited Partnership | System and method for integration of communication systems with computer-based information systems |
US6101473A (en) * | 1997-08-08 | 2000-08-08 | Board Of Trustees, Leland Stanford Jr., University | Using speech recognition to access the internet, including access via a telephone |
US6801763B2 (en) * | 1997-10-29 | 2004-10-05 | Metro One Telecommunications, Inc. | Technique for effectively communicating travel directions |
US6104788A (en) * | 1997-12-04 | 2000-08-15 | Siemens Information And Communication Networks, Inc. | Apparatus and method for using a telephone for remote scheduling |
US6704394B1 (en) * | 1998-03-25 | 2004-03-09 | International Business Machines Corporation | System and method for accessing voice mail from a remote server |
US6640242B1 (en) * | 1999-01-29 | 2003-10-28 | Microsoft Corporation | Voice access through a data-centric network to an integrated message storage and retrieval system |
US6411685B1 (en) * | 1999-01-29 | 2002-06-25 | Microsoft Corporation | System and method for providing unified messaging to a user with a thin web browser |
US6795536B1 (en) * | 1999-02-26 | 2004-09-21 | Mitel, Inc. | Automatic user preference selection for message playback based on caller line identification data |
US6981023B1 (en) * | 1999-03-09 | 2005-12-27 | Michael Hamilton | Message routing |
US7120234B1 (en) * | 1999-12-29 | 2006-10-10 | Bellsouth Intellectual Property Corp. | Integrated tone-based and voice-based telephone user interface |
US20010014590A1 (en) * | 2000-01-19 | 2001-08-16 | Akihiro Nakano | Portable telephone apparatus and stablly supplying method of reference frequency |
US7149777B1 (en) * | 2000-01-20 | 2006-12-12 | Avaya Technology Corp. | Method for administering message attachments |
US20010015972A1 (en) * | 2000-02-21 | 2001-08-23 | Shoichi Horiguchi | Information distributing method, information distributing system, information distributing server, mobile communication network system and communication service providing method |
US6547002B1 (en) * | 2000-04-17 | 2003-04-15 | Weatherford/Lamb, Inc. | High pressure rotating drilling head assembly with hydraulically removable packer |
US6920425B1 (en) * | 2000-05-16 | 2005-07-19 | Nortel Networks Limited | Visual interactive response system and method translated from interactive voice response for telephone utility |
US6651042B1 (en) * | 2000-06-02 | 2003-11-18 | International Business Machines Corporation | System and method for automatic voice message processing |
US7080315B1 (en) * | 2000-06-28 | 2006-07-18 | International Business Machines Corporation | Method and apparatus for coupling a visual browser to a voice browser |
US6697458B1 (en) * | 2000-07-10 | 2004-02-24 | Ulysses Esd, Inc. | System and method for synchronizing voice mailbox with e-mail box |
US6738462B1 (en) * | 2000-07-19 | 2004-05-18 | Avaya Technology Corp. | Unified communications automated personal name addressing |
US20020055350A1 (en) * | 2000-07-20 | 2002-05-09 | Ash Gupte | Apparatus and method of toggling between text messages and voice messages with a wireless communication device |
US20030046401A1 (en) * | 2000-10-16 | 2003-03-06 | Abbott Kenneth H. | Dynamically determing appropriate computer user interfaces |
US7068643B1 (en) * | 2000-11-03 | 2006-06-27 | Intervoice Limited Partnership | Extensible interactive voice response |
US20020082030A1 (en) * | 2000-11-29 | 2002-06-27 | Stefan Berndt | Apparatus and method for forwarding messages to terminals of various communication media |
US20020077829A1 (en) * | 2000-12-19 | 2002-06-20 | Brennan Paul Michael | Speech based status and control user interface customisable by the user |
US7283808B2 (en) * | 2001-01-18 | 2007-10-16 | Research In Motion Limited | System, method and mobile device for remote control of a voice mail system |
US20020109731A1 (en) * | 2001-02-10 | 2002-08-15 | International Business Machines Corporation | Facilitated user interface |
US7054939B2 (en) * | 2001-06-28 | 2006-05-30 | Bellsouth Intellectual Property Corportion | Simultaneous visual and telephonic access to interactive information delivery |
US20030046346A1 (en) * | 2001-07-11 | 2003-03-06 | Kirusa, Inc. | Synchronization among plural browsers |
US20030059000A1 (en) * | 2001-09-27 | 2003-03-27 | Burton David Kearney | Integrated voice access to a variety of personal information services |
US7289606B2 (en) * | 2001-10-01 | 2007-10-30 | Sandeep Sibal | Mode-swapping in multi-modal telephonic applications |
US7356130B2 (en) * | 2001-10-11 | 2008-04-08 | International Business Machines Corporation | Method and system for selecting speech or DTMF interfaces or a mixture of both |
US6987840B1 (en) * | 2001-11-06 | 2006-01-17 | At&T Corp. | Integrated message management method and system |
US6778644B1 (en) * | 2001-12-28 | 2004-08-17 | Vocada, Inc. | Integration of voice messaging and data systems |
US20030235282A1 (en) * | 2002-02-11 | 2003-12-25 | Sichelman Ted M. | Automated transportation call-taking system |
US20030182622A1 (en) * | 2002-02-18 | 2003-09-25 | Sandeep Sibal | Technique for synchronizing visual and voice browsers to enable multi-modal browsing |
US20030162561A1 (en) * | 2002-02-27 | 2003-08-28 | Greg Johnson | System and method for concurrent multimodal communication session persistence |
US7315613B2 (en) * | 2002-03-11 | 2008-01-01 | International Business Machines Corporation | Multi-modal messaging |
US20030195937A1 (en) * | 2002-04-16 | 2003-10-16 | Kontact Software Inc. | Intelligent message screening |
US20040148356A1 (en) * | 2002-11-04 | 2004-07-29 | Bishop James William | System and method for private messaging |
US20040141594A1 (en) * | 2003-01-20 | 2004-07-22 | Brunson Gordon R. | Messaging advise in presence-aware networks |
US20040156484A1 (en) * | 2003-02-12 | 2004-08-12 | Umesh Amin | Interface for a multi-messaging system |
US20050054381A1 (en) * | 2003-09-05 | 2005-03-10 | Samsung Electronics Co., Ltd. | Proactive user interface |
US7136478B1 (en) * | 2004-01-13 | 2006-11-14 | Avaya Technology Corp. | Interactive voice response unit response display |
US20060083358A1 (en) * | 2004-10-20 | 2006-04-20 | Microsoft Corporation | Unified messaging architecture |
US20060083357A1 (en) * | 2004-10-20 | 2006-04-20 | Microsoft Corporation | Selectable state machine user interface system |
US7551727B2 (en) * | 2004-10-20 | 2009-06-23 | Microsoft Corporation | Unified messaging architecture |
US20090290692A1 (en) * | 2004-10-20 | 2009-11-26 | Microsoft Corporation | Unified Messaging Architecture |
US7912186B2 (en) * | 2004-10-20 | 2011-03-22 | Microsoft Corporation | Selectable state machine user interface system |
US8090083B2 (en) * | 2004-10-20 | 2012-01-03 | Microsoft Corporation | Unified messaging architecture |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120194877A1 (en) * | 2003-02-11 | 2012-08-02 | Omtool, Inc. | Method and system for secure facsimile delivery and registration |
US8630011B2 (en) * | 2003-02-11 | 2014-01-14 | Omtool, Ltd. | Method and system for secure facsimile delivery and registration |
US8090083B2 (en) | 2004-10-20 | 2012-01-03 | Microsoft Corporation | Unified messaging architecture |
US8904270B2 (en) | 2006-11-29 | 2014-12-02 | Omtool Ltd. | Methods and apparatus for enterprise document distribution |
US20120281819A1 (en) * | 2011-05-05 | 2012-11-08 | Shreedhar Madhavapeddi | Voicemail preview and editing system |
US8913722B2 (en) * | 2011-05-05 | 2014-12-16 | Nuance Communications, Inc. | Voicemail preview and editing system |
US9497327B2 (en) | 2011-05-05 | 2016-11-15 | Nuance Communications, Inc. | Voicemail preview and editing system |
US11212384B2 (en) | 2019-12-20 | 2021-12-28 | Capital One Services, Llc | Cross-channel configurable navigation |
US11671529B2 (en) | 2019-12-20 | 2023-06-06 | Capital One Services, Llc | Cross-channel configurable navigation |
Also Published As
Publication number | Publication date |
---|---|
JP2006268832A (en) | 2006-10-05 |
US7912186B2 (en) | 2011-03-22 |
CN1838694A (en) | 2006-09-27 |
US20060083357A1 (en) | 2006-04-20 |
KR20060102268A (en) | 2006-09-27 |
EP1705886A1 (en) | 2006-09-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7912186B2 (en) | Selectable state machine user interface system | |
US8090083B2 (en) | Unified messaging architecture | |
US7027463B2 (en) | System and method for multi-tiered rule filtering | |
US7484213B2 (en) | Agent architecture employed within an integrated message, document and communication system | |
KR101691239B1 (en) | Enhanced voicemail usage through automatic voicemail preview | |
US6781962B1 (en) | Apparatus and method for voice message control | |
JP4607493B2 (en) | Enhanced telephony computer user interface | |
JP5615922B2 (en) | Mashups and presence found on the phone | |
US8504633B2 (en) | Method and system for communicating a data file | |
US9712672B2 (en) | Call response system | |
US8706092B2 (en) | Outgoing voice mail recording and playback | |
US20060031340A1 (en) | Apparatus and method for advanced attachment filtering within an integrated messaging platform | |
US20040252679A1 (en) | Stored voice message control extensions | |
US8706091B2 (en) | Attachment of rich content to a unified message left as a voicemail | |
JP2008061241A (en) | Method and communication system for continuously recording surrounding information | |
WO2003073678A2 (en) | Method and apparatus for switching between a circuit switched channel and a packet data network channel |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |
|
AS | Assignment |
Owner name: MICROSOFT TECHNOLOGY LICENSING, LLC, WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICROSOFT CORPORATION;REEL/FRAME:034766/0001 Effective date: 20141014 |