[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

AU2016246072B2 - Visual interactive voice response system - Google Patents

Visual interactive voice response system Download PDF

Info

Publication number
AU2016246072B2
AU2016246072B2 AU2016246072A AU2016246072A AU2016246072B2 AU 2016246072 B2 AU2016246072 B2 AU 2016246072B2 AU 2016246072 A AU2016246072 A AU 2016246072A AU 2016246072 A AU2016246072 A AU 2016246072A AU 2016246072 B2 AU2016246072 B2 AU 2016246072B2
Authority
AU
Australia
Prior art keywords
interaction
communications device
voice
instructions
site
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
AU2016246072A
Other versions
AU2016246072A1 (en
Inventor
Andrea Friio
Scott Kolman
Yochai Konig
Praphul Kumar
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Genesys Cloud Services Holdings II LLC
Original Assignee
Genesys Cloud Services Holdings II LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Genesys Cloud Services Holdings II LLC filed Critical Genesys Cloud Services Holdings II LLC
Publication of AU2016246072A1 publication Critical patent/AU2016246072A1/en
Application granted granted Critical
Publication of AU2016246072B2 publication Critical patent/AU2016246072B2/en
Assigned to Genesys Cloud Services Holdings II, LLC reassignment Genesys Cloud Services Holdings II, LLC Request to Amend Deed and Register Assignors: GREENEDEN U.S. HOLDINGS II LLC
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M3/00Automatic or semi-automatic exchanges
    • H04M3/42Systems providing special services or facilities to subscribers
    • H04M3/487Arrangements for providing information services, e.g. recorded voice services or time announcements
    • H04M3/493Interactive information services, e.g. directory enquiries ; Arrangements therefor, e.g. interactive voice response [IVR] systems or voice portals
    • H04M3/4938Interactive information services, e.g. directory enquiries ; Arrangements therefor, e.g. interactive voice response [IVR] systems or voice portals comprising a voice browser which renders and interprets, e.g. VoiceXML
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/10Architectures or entities
    • H04L65/1063Application servers providing network services
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/02Protocols based on web technology, e.g. hypertext transfer protocol [HTTP]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2203/00Aspects of automatic or semi-automatic exchanges
    • H04M2203/25Aspects of automatic or semi-automatic exchanges related to user interface aspects of the telephonic communication service
    • H04M2203/251Aspects of automatic or semi-automatic exchanges related to user interface aspects of the telephonic communication service where a voice mode or a visual mode can be used interchangeably
    • H04M2203/253Aspects of automatic or semi-automatic exchanges related to user interface aspects of the telephonic communication service where a voice mode or a visual mode can be used interchangeably where a visual mode is used instead of a voice mode
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M3/00Automatic or semi-automatic exchanges
    • H04M3/42Systems providing special services or facilities to subscribers
    • H04M3/42136Administration or customisation of services

Landscapes

  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Telephonic Communication Services (AREA)
  • Information Transfer Between Computers (AREA)

Abstract

A request is received from a communications device to execute an interaction site. A request is transmitted to the automated response system. First instructions that provide one or more steps of the multi-step communication flow between the communications device and the automated response system are received from the automated response system. In response to determining that the request is for the voice-interaction with the interaction site, second instructions that provide the one or more steps through a voice interaction with a user of the communications device are determined and transmitted to the communications device. In response to determining that the request is for the visual-interaction with the interaction site, third instructions that provide the one or more steps through a visual interaction with the user of the communications device are determined and transmitted to the communications device.

Description

VISUAL INTERACTIVE VOICE RESPONSE SYSTEM
BACKGROUND
The following disclosure relates generally to the development and delivery of a multichannel interaction application. A user may use a personal communications device to 5 contact a company to discuss a service or a product provided by the company. The user may contact the company by establishing electronic communications with the company over a voice channel or a data channel.
Throughout this specification the word comprise, or variations such as comprises or comprising, will be understood to imply the inclusion of a stated element, integer or 10 step, or group of elements, integers or steps, but not the exclusion of any other element, integer or step, or group of elements, integers or steps.
Any discussion of documents, acts, materials, devices, articles or the like which has been included in the present specification is not to be taken as an admission that any or all of these matters form part of the prior art base or were common general 15 knowledge in the field relevant to the present disclosure as it existed before the priority date of each of the appended claims.
SUMMARY
In a general aspect, a computer-implemented method, comprises a request is received by an application server and from a communications device to execute an interaction 20 site for enabling communications to be exchanged between an automated response system and the communications device, where the interaction site specifies a multistep communication flow between the communications device and the automated response system. A request to execute the interaction site is transmitted from the application server and to the automated response system. A common interaction flow 25 document including first instructions that, when executed, provide one or more steps of the multi-step communication flow between the communications device and the
-12016246072 28 Mar 2019 automated response system, wherein the common interaction flow document is for handling interactions received over different channels including voice interactions and visual interactions are received from the automated response system by the application server. Whether the request from the communications device is for a voice-interaction 5 with the interaction site or a visual-interaction with the interaction site is determined by the application server. In response to determining that the request from the communications device is for the voice-interaction with the interaction site, retrieving the common interaction flow document; wherein the second instructions, when executed by the communications device, provide the one or more steps through a 10 voice interaction with a user of the communications device are translated by the application server, the first instructions into second instructions. The second instructions are transmitted from the application server to the communications device. In response to determining that the request from the communications device is for the visual-interaction with the interaction site, retrieving the common interaction flow 15 document; wherein the third instructions that, when executed by the communications device, provide the one or more steps through a visual interaction with the user of the communications device are translated by the application server, the first instructions into third instructions. The third instructions are transmitted from the application server to the communications device.
Implementations may include one or more of the following features. To transmit, from the application server and to the automated response system, the request to execute the interaction site, an indication that indicates whether the request from the communications device is for the voice-interaction with the interaction site or the visualinteraction with interaction site may be transmitted to the automated response system.
To determine whether the request from the communications device is for the voiceinteraction with the interaction site or the visual-interaction with the interaction site, whether the request is received from a voice channel or a data channel may be determined. In response to determining that the request is received from the voice
-22016246072 28 Mar 2019 channel, the request is for the voice-interaction with the interaction site may be determined. In response to determining that the request is received from the data channel, the request is for the visual-interaction with the interaction site may be determined.
To determine whether the request from the communications device is for the voiceinteraction with the interaction site or the visual-interaction with the interaction site, whether the request is based on hypertext markup language (HTML) may be determined. In response to determining that the request is based on HTML, the request from the communications device is for the visual-interaction with the interaction 10 site may be determined.
Data representing a communication between the communications device and the automated response system may be stored, where the data may include an indication of whether the communication is visual. The interaction site may be associated with one or more interaction pages, where each interaction page may represent a 15 respective step of the multi-step communication flow, and where the one or more voice interaction may be configured by a user of a content provider device through interactions with a development interface to include values for (i) one or more parameters associated with the voice-interaction with the interaction site, (ii) one or more parameters associated with the visual-interaction with the interaction site, and (iii) 20 one or more parameters associated with both the voice-interaction and the visualinteraction with the interaction site. The second instructions may not include instructions generated using the values for the one or more parameters associated with the visual-interaction with the interaction site. The third instructions may not include instructions generated using the values for the one or more parameters associated with 25 the voice-interaction with the interaction site.
The first instructions may be based on state chart extensible markup language (SCXML) or VoiceXML. The second instructions may based on voice extensible
-32016246072 28 Mar 2019 markup language (VoiceXML). The third instructions may be based on HTML5. The automated response system may be an interactive voice response system.
In another general aspect, a system includes one or more processors and one or more non-transitory computer-readable storage devices storing instructions that, when 5 executed by the one or more processors, cause the one or more processors to perform operations including receiving, from a communications device, a request to execute an interaction site for enabling communications to be exchanged between an automated response system and the communications device, where the interaction site specifies a multi-step communication flow between the communications device and the automated 10 response system; transmitting, to the automated response system, a request to execute the interaction site; receiving, from the automated response system, a common interaction flow document including first instructions that, when executed, provide one or more steps of the multi-step communication flow between the communications device and the automated response system, wherein the common 15 interaction flow document is for handling interactions received over different channels including voice interactions and visual interactions; determining whether the request from the communications device is for a voice-interaction with the interaction site or a visual-interaction with the interaction site; in response to determining that the request from the communications device is for the voice-interaction with the interaction site: 20 retrieving the common interaction flow document; translating the first instructions into second instructions, wherein the second instructions, when executed by the communications device, provide the one or more steps through a voice interaction with a user of the communications device; and transmitting, to the communications device, the second instructions; and in response to determining that the request from the 25 communications device is for the visual-interaction with the interaction site: retrieving the common interaction flow document; translating the first instructions into third instructions, wherein the third instructions, when executed by the communications device, provide the one or more steps through a visual interaction with the user of the
-42016246072 28 Mar 2019 communications device; and transmitting, to the communications device, the third instructions.
A non-transitory computer-readable storage device storing instructions that when executed by one or more processors, cause the one or more processors to perform 5 operations comprising: receiving, from a communications device, a request to execute an interaction site for enabling communications to be exchanged between an automated response system and the communications device, wherein the interaction site specifies a multi-step communication flow between the communications device and the automated response system; transmitting, to the automated response system, a 10 request to execute the interaction site; receiving, from the automated response system, first instructions that, when executed, provide one or more steps of the multi-step communication flow between the communications device and the automated response system; determining whether the request from the communications device is for a voice-interaction with the interaction site or a visual-interaction with the interaction site; 15 in response to determining that the request from the communications device is for the voice-interaction with the interaction site: retrieving the common interaction flow document; translating the first instructions into second instructions, wherein the second instructions, when executed by the communications device, provide the one or more steps through a voice interaction with a user of the communications device; and 20 transmitting, to the communications device, the second instructions; and in response to determining that the request from the communications device is for the visualinteraction with the interaction site: retrieving the common interaction flow document; translating the first instructions into third instructions, wherein the third instructions, when executed by the communications device, provide the one or more steps through 25 a visual interaction with the user of the communications device; and, transmitting, to the communications device, the third instructions.
-52016246072 28 Mar 2019
The details of one or more implementations are set forth in the accompanying drawings and the description below. Other potential features and advantages will become apparent from the description, the drawings, and the claims.
BRIEF DESCRIPTION OF THE FIGURES
FIG. 1 is a block diagram of a communications system that provides development and access of visual-IVR enabled interaction sites.
FIGS. 2A-2C illustrate a GUI for an application development tool that is used by a content provider to create a visual-IVR enabled interaction site.
FIGS. 3A-3B illustrate a GUI for an application development tool that is used by a 10 content provider to create a visual-IVR enabled message page.
FIG. 3C illustrates a rendered message page in a visual-IVR application.
FIGS. 4A-4B, 5A-5B, and 6A-6B illustrate a GUI for an application development tool that is used by a content provider to create visual-IVR enabled question pages.
FIGS. 4C, 5C, 6C, and 7-10 illustrate rendered question pages in a visual-IVR 15 application.
FIG. 11 is a flow chart illustrating an example of a process for a content provider to create and host a visual-IVR enabled interaction site.
FIG. 12 is a flow chart illustrating an example of a process for a user to communicate with a call handling system using a visual-IVR enabled voice site.
DETAILED DESCRIPTION
A company may wish to enable its customers to contact it through use of any of multiple different communications channels (e.g., phone, email, chat, SMS or another communications channel that support communications between a customer and a service/product provider). To save costs, the company may wish to interact with the 25 customers, at least initially, using an automated response system. Use of such an automated response system may decrease the costs associated with responding to customer contacts by decreasing the number of human agents that need to be
-62016246072 28 Mar 2019 employed to respond to the customer contacts. Many customer contacts are for routine inquiries (e.g., a customer contacts the company to check the balance of a checking account or of a credit card) and, therefore, lend themselves well to being efficiently and cost-effectively handled through use of an automated response system.
In contrast, more complex customer contacts, which are typically far fewer in number, may be routed to human agents for handling.
The interaction between a customer and an automated response system may be designed by a content provider offering the service using an interaction flow that segments the interaction into multiple discrete steps. For example, when a customer 10 dials a company’s telephone number to pay a bill, a voice interaction flow may include the steps of (i) acquiring the customer’s account number, (ii) informing the customer of a current balance, (iii) acquiring the customer’s payment information, (iv) processing payment, and (v) conducting a customer satisfaction survey. An interactive voice response system (IVR) is an example automated response system that supports a 15 voice-based interaction between a user operating a telephonic device and a third party automated application, where the user interacts with the third party automated application following the multi-step interaction flow that both provides and collects information to and from the user using voice communications with the IVR.
Many advanced telephonic devices (e.g., smart phones) include functionalities that 20 support visual interactions with users. For example, a smart phone may display information to a user on a display using a scripting language such as HTML5. As another example, a touchscreen of a smart phone may allow a user to navigate through content by touching the touchscreen. As another example, a smart phone may allow a user to input text information as an alternative of using voice. Given that some 25 users may wish to visually interact with a third party automated application instead of using voice, having an automated response system that can enable both voice and visual interactions with users using a same interaction flow may be useful. A visual IVR (VIVR) system may be implemented to process a multi-step interaction according
-72016246072 28 Mar 2019 to its current state, where the processing is independent of whether the interaction is a visual interaction or a voice interaction. Notably, a visual IVR is an ARS that is able to support voice only and visual only applications, and also a mixed voice-visual application.
More specifically, a user of a particular product or service provided by a company may need to contact customer service for the product or service to troubleshoot a problem the user is experiencing in using the product or service. In order to contact the customer service and obtain a solution to the problem, the user may use a smart phone to call a known customer service number for the product or service. By calling 10 the customer service number, the user may get connected to a call handling system that enables the user to have a voice interaction with a voice site associated with the product or service according to an interaction flow. Alternatively, the user may initiate a visual interaction with the voice site using a client-side application running on the smart phone, where the voice site may be accessed by entering a uniform resource 15 identifier (URL) in the client-side application. By entering the URL, the user may get connected to the call handling system that enables the user to have a visual interaction with the voice site associated with the product or service according to the same interaction flow. End user has option to go through automated service either through voice or mobile web channel or both at the same time.
A voice site is a set of scripts or, more generally, programming language modules corresponding to one or more linked pages that collectively interoperate to produce an automated interactive experience with a user. A standard voice site includes scripts or programming language modules corresponding to at least one voice page and limits the interaction with the user to an audio communications mode. Because customers 25 typically access a voice site by calling a telephone number using a telephone, a standard voice site is typically referred to as a single channel interaction site, i.e., an interaction site that supports a single type of contact. A VIVR enables customers to
-82016246072 28 Mar 2019 visually interact with the voice site by entering a URL on a telephonic device, and thus the standard voice site may be expanded into a multi-channel interaction site.
The voice site accessed by the user may be an automated interactive voice site that is configured to process, using pre-programmed scripts, information received from the 5 user that is input through the telephonic device being used by the user, and, in response, provide information to the user that is conveyed to the user through the telephonic device. The interaction between the user and the voice site may be done using a VIVR provided by a service provider that is hosting the voice site and an application server that acts as a gateway between the telephonic device and the VIVR.
For either a voice interaction or a visual interaction, the VIVR may be configured to support commands and information using scripts that are pre-programmed for the voice site, for example, voice-extensible markup language (VoiceXML) scripts or state chart XML (SCXML) scripts. In a voice interaction, the VIVR interacts with the user by using audible commands to prompt the user to provide information and enabling the 15 user to input the information by speaking into the telephonic device or by pressing buttons on the telephonic device (when using, for example, a touch-tone telephone).
The information input by the user is conveyed to the VIVR over a voice communications session that is established between the telephonic device and the IVR when the call is connected, with the application server as a gateway. Upon receiving 20 the information, the VIVR processes the information using the pre-programmed scripts.
The VIVR may be configured to send audible responses back to the user via the telephonic device.
In a visual interaction, scripts that are pre-programmed for the voice site may be translated by an application server into a different programming or scripting language 25 (e.g., HTML5) such that the translated scripts may be visually presented to the user using a client-side application running on the telephonic device. The VIVR interacts with the user by using visual commands to prompt the user to provide information and enabling the user to input the information by entering text and numbers using the
-92016246072 28 Mar 2019 telephonic device. The information input by the user is translated by the application server and conveyed to the VIVR over a data communications session that is established between the telephonic device and the VIVR, with the application server as a gateway. Upon receiving the information, the VIVR processes the information using 5 the pre-programmed scripts. Notably, the processing of the information by the VIVR may be the same regardless whether it is a voice interaction or a visual interaction. This also allows customers to switch from one channel, say voice, to another channel (mobile web like HTML5) without losing context.
Distinguishable from a visual presentation of a standard voice site, an enhanced voice 10 site may include scripts or programming language modules corresponding to at least one voice page and at least one multimodal action page linked to the at least one voice page that enable interaction with the user to occur via an audio communications mode and at least one additional communications mode (e.g., a text communications mode, an image communications mode or a video communications mode). For example, a 15 customer may call a contact center from a smart phone and may be placed on hold in a queue. If the wait time exceeds a threshold, the voice site may offer a hyperlink visually to the customer via the smart phone, where the customer may click on the hyperlink to schedule a call back. As another example, a customer may call a contact center to schedule an appointment, where a voice site may enable visual interaction 20 with the user by showing a calendar for making the appointment. As another example, during a voice call with a customer, an agent may offer visual assistance to the customer by pushing through SMS, email, or a push notification of a session to a calendar on the customer’s smart phone. As another example, upon completion of an interaction with a voice site, a customer may receive a call back from the interactive 25 voice response system for a survey of the experience, where a link may be provided to the customer via the smart phone to allow the customer to visually rate the experience with the voice site. An enhanced voice site may, therefore, be referred to as a voice site that has been enhanced to enable some portions of the interaction flow to involve
-102016246072 28 Mar 2019 the communication of multimedia information. In some implementations, an enhanced voice site may be configured to support multimedia information including audio, video, images and text. Notably, the voice pages in an enhanced voice site may be accessed as either a voice interaction or a visual interaction using VIVR.
A voice site may be hosted by a third party service provider that facilitates the creation and hosting of voice sites, or more generally, interaction sites on servers owned and operated by the service provider. An interaction site may be a set of scripts or programming modules that offer a common interaction flow for handling contacts received over different channels. The set of scripts or programming modules may then 10 be translated by an interaction flow processor into a corresponding set of channelspecific scripts or programming modules for each channel supported by the interaction site, and these translated channel-specific scripts or programming modules may then be executed by the respective sub-systems of the multi-channel communications system to enable automated interactions with users over the different channels. For 15 example, the pre-programmed scripts of the interaction site may be extensible markup language (XML) scripts. If the user accesses the multi-channel communications system by using a telephone to call a telephone number associated with the interaction site, the interaction flow processor may translate the XML scripts of the interaction site to VoiceXML scripts for processing by an IVR to interact with the calling user.
The service provider may provide a service/method that enables the design, development, and hosting of VIVR-enabled interaction sites or applications that run a thin client on the communications device that interacts with a fully hosted, on-demand interaction solution platform maintained and managed by the service provider. The service/method provides a way to develop VIVR-enabled voice site that is supported by 25 a VIVR system (the server side) and allows a voice or visual interaction between the client and the server to be established. In some implementations, the service/method may require an installation of a thin client engine (e.g., an application) on the communications device of the user that mediates between the objects and devices in
-112016246072 28 Mar 2019 the communications device and the multi-channel communications system supporting the interaction site hosted by the service provider.
In the above scenario, the role of the entity providing customer service through the interaction site is that of a content provider. The developer of the entity/company 5 (hereinafter referred to interchangeably as the “content provider”) configures the interaction site that is to be used for the particular product or service and provides the logic for the interaction site that is to be executed by the multi-channel communications system. The content provider may do so by using a graphical user interface (GUI) provided by the third party service provider for configuring the interaction site. The 10 service provider handles the interpretation and compilation of the information provided by the content provider, and the creation and hosting of the interaction site based on the information. Since the service provider manages the multi-channel communications system, the service provider may allow the content provider to develop the interaction site using one unified GUI interface, where the interaction site is 15 executable by the user via any one or more of multiple different communications channels.
The service/method thus enables the deployment of interaction-enabled solutions on communications devices without requiring the content provider to engage in complex programming. Applications, or interaction sites, may be designed by the content 20 provider using a web-based or remotely accessible interface, and served on demand to clients. In some implementations, clients can be add-ons that smart phone applications can plug into. In some implementations, the service/method enable users to interact with a VIVR application. The VIVR application enables users to contact and interact with a VIVR-enabled voice site via a voice channel or a data channel.
FIG. 1 is a block diagram of a communications system 100 that provides a development platform for VIVR-enabled interaction sites and access to these VIVRenabled interaction sites in a multi-channel solution platform. Referring to Fig. 1, a user of a communications device (i.e., a smart phone) 110 is able to interact with the
-122016246072 28 Mar 2019 communications device 110 to request a service from a VIVR-enabled interaction site that is provided, for example, by a content provider. The service may be, for example, a request to purchase a particular product or service offered by or made available by the content provider through the interaction site. For example, the user may indicate a 5 desire to request a service from the interaction site by selecting a graphically displayed icon on a graphical user interface (GUI) of the communications device 110 to thereby invoke an application stored in the communications device 110 with which the user can interact to initiate a service request. Additionally or alternatively, the user may indicate a desire to request a service by inputting, via manual selection or otherwise, a 10 telephone number associated with the customer service department into the communications device 110 and initiating a call directed to the inputted telephone number. Additionally or alternatively, the user may indicate a desire to request a service via a communications channel not listed in the above examples.
In some implementations, the request for the service or product may be directed to an 15 application server 140 of a handling system 150. The application server 140 may forward the request to a VIVR system 152 of the call handling system 150, and an interaction site may be invoked, where the VIVR system 152 communicates with the communications device 110 through the application server 140 to provide the requested service. As mentioned previously, an interaction site may be hosted by a 20 third party service provider that facilitates the creation and hosting of interaction sites on servers owned and operated by the service provider. Depending on the communications means that a user uses to access the VIVR-enabled interaction site, the application server 140 may process the communications between the VIVR system 152 and the communications device 110 according to the communications means to 25 process the request. For example, a call handling system may present the interaction site to the user using voice messages that are generated by VoiceXML scripts. If the interaction is a visual one, the application server 140 may translate the voice messages into corresponding messages in a different scripting or programming
-132016246072 28 Mar 2019 language (e.g., HTML5), so the corresponding messages may be displayed on the communications device 110. It also allows to offer additional contents like graphics, video as well as audio to the users which are not supported over IVR. This not only complements the IVR functionality as well as extends it.
In many applications, the flow for providing a service to the user includes the same steps regardless of which communications form the user is using. From a content provider’s perspective, it is a burden to require developing a VIVR-enabled interaction site for each of the communications forms (e.g., a voice interaction or a visual interaction) using different tools or scripting languages. From a service provider’s 10 perspective, the storage and management of a VIVR-enabled interaction site having different versions for each of the communications form may be complicated. Accordingly, a communications system that can integrate the development of a VIVRenabled interaction site for each of the communications forms using one development platform, and compile the developed interaction site into one scripting language that 15 can be translated based on the communications channel used by a user may enable a content provider and/or service provider to enjoy a decrease in costs associated with developing and managing interaction sites without compromising the quality of the user experience with the VIVR-enabled interaction site.
The communications system 100 is an example implementation of a system that 20 supports an interactive VIVR application delivery platform. In general, the communications system 100 includes the communications device 110, a network 130, the call handling system 150, a content provider system 142, a data store 160, an interaction flow processor 180, an interaction flow document database 185, and an application builder 190.
The communications device 110 is configured to allow a user to interact with the call handling system 150 across the network 130. The communications device 110 may be a cellular phone or a mobile personal digital assistant (PDA) with embedded cellular phone technology. The communications device 110 may be a computer that includes
-142016246072 28 Mar 2019 one or more software or hardware applications for performing communications between the communications device 110 and the multi-channel communications system 135. For example, a web browser may be installed on the communications device 110 to allow the user to access a VIVR-enabled voice site using a URL. As 5 another example, a mobile application that is developed by a content provider may be installed on the communications device 110 to allow the user to access a VIVRenabled voice site of the content provider directly using the mobile application. As another example, a mobile application that is developed by a service provider may be installed on the communications device 110 to allow the user to access different VIVR10 enabled voice sites of different content providers using the mobile application. The communications device 110 may have various input/output devices with which a user may interact to provide and receive audio, text, video, and other forms of data.
The network 130 may include a telephone network, a data network, or a combination of the telephone network and the data network. The telephone network may include a 15 circuit-switched voice network, a packet-switched data network, or any other network able to carry voice data. For example, circuit-switched voice networks may include a Public Switched Telephone Network (PSTN), and packet-switched data networks may include networks based on the Internet protocol (IP) or asynchronous transfer mode (ATM), and may support voice using, for example, Voice-over-IP, Voice-over-ATM, or 20 other comparable protocols used for voice data communications. The data network is configured to enable direct or indirect communications between the communications device 110 and the call handling system 150. Examples of the data network include the Internet, Wide Area Networks (WANs), Local Area Networks (LANs), analog or digital wired and wireless telephone networks (e.g., Public Switched Telephone 25 Network (PSTN), Integrated Services Digital Network (ISDN), and Digital Subscriber
Line (xDSL)), radio, television, cable, satellite, and/or any other delivery or tunneling mechanism for carrying data. In some implementations, the data network and the telephone network are implemented by a single or otherwise integrated
-152016246072 28 Mar 2019 communications network configured to enable communications between the communications device 110 and the call handling system 150.
A call handling system 150 receives the request and interacts with the communications device 110 to provide the requested service through the VIVR-enabled voice site. The 5 call handling system 150 may include a VIVR system 152 and an application server 140. In some implementations, the call handling system 150 may additionally include an agent routing system 154.
The VIVR system 152 may include a voice gateway coupled to an interaction flow processor 180 via the network 130. Alternatively, the voice gateway may be local to 10 the interaction flow processor 180 and connected directly to the interaction flow processor 180. The voice gateway is a gateway that receives user calls from or places calls to voice communications devices, such as the communications device 110, and responds to the calls in accordance with a voice program that corresponds to a flow of an interaction site. The voice program may be accessed from local memory within the 15 voice gateway or from the interaction flow processor 180. In some implementations, the voice gateway processes voice programs that are script-based voice applications.
The voice program, therefore, may be a script written in a scripting language such as, for example, voice extensible markup language (VoiceXML), SCXML, or speech application language tags (SALT). The VIVR system 152 may also be configured to 20 communicate with the data store 160 to read and/or write user interaction data (e.g., state variables for a data communications session) in a shared memory space.
The application server 140 is configured to establish a voice or a visual communications session between the communications device 110 and the VIVR system 152, and to receive and send data to the communications device 110 across 25 the network 130. The application server 140 is configured to communicate with the
VIVR system 152 to send data received from the communications device 110. The application server 140 also may send other application-related data that did not originate from the communications device 110 to the VIVR system 152. For example,
-162016246072 28 Mar 2019 the application server 140 may send to the VIVR system 152 data indicating whether the communications session is voice or visual. In some implementations, the application server 140 is configured to communicate with the data store 160 to read and/or write user interaction data (e.g., state variables for a data communications session) in a shared memory space. As shown in Fig. 1, the application server 140 and the call handling system 150 are a single integrated computer system. In some other implementations, the application server 140 may be one or more computer systems that operate separately or in concert under the direction of one or more software programs to perform the above-noted functions.
The agent routing system 154 of the call handling system may include, among other components, an inbound call queue, an outbound call request queue, a call router, an automatic call distributor (ACD) administrator, and contact information of a plurality of agents. The agent routing system 154 may receive one or more calls from one or more voice communication devices, such as the communications device 110, via the network 15 130 and may make one or more outbound calls to voice communication devices of an agent via the network 130. The agent routing system 154 may determine an appropriate agent to route the call to or to assign an outbound call to. The determination of an appropriate agent may be based on agent skills, agent performance metrics and information known about the inbound or outbound call and the status of the agent.
The interaction flow processor 180 includes all hardware and software components that interface and provide data to the call handling system 150. In some implementations, the interaction flow processor 180 sends translated application programs or scripts to the call handling system 150 for processing user interactions 25 with a voice site. The user interactions are analyzed by the call handling system 150 and new programs or scripts that correspond to the next state of the interaction flow may then be sent to the call handling system 150 for further processing. In some implementations, the interaction flow processor 180 may determine which programs or
-172016246072 28 Mar 2019 scripts to provide to the call handling system 150 based on some or all of the information received from the call handling system 150 or the communications device 110.
The interaction flow document database 185 stores interaction flow documents created by the application builder 190, and provides the interaction flow processor 180 access to these interaction flow documents. In some implementations, the interaction flow document database 185 may be an array of high-capacity storage drives that are closely coupled to the interaction flow processor 180. In some implementations, the interaction flow document database 185 may be an array of high-capacity storage 10 drives that are closely coupled to the application builder 190.
The content provider system 142 is configured to allow a content provider to interact with the application builder 190 across the network 130. The content provider system 142 may be a cellular phone or a mobile personal digital assistant (PDA) with embedded cellular phone technology. The content provider system 142 may be a 15 computer that includes one or more software or hardware applications for performing communications between content provider system 142 and the application builder 190.
The content provider system 142 may have various input/output devices with which a content provider may interact to provide and receive audio, text, video, and other forms of data from the application builder 190.
The application builder 190 facilitates the creation of interaction sites that include voice sites. The application builder 190 utilizes various components to enable the creation of interaction sites. The various components of the application builder 190 may be colocated in a single physical location, or they may be geographically distributed, with dedicated high capacity links interconnecting the various components. The application builder 190 may include a content provider interface 192 and an application compiler 194.
The content provider interface 192 is a GUI front-end for an application development tool that can be used to build an interaction site that is capable of handling interactions
-182016246072 28 Mar 2019 using multiple communications channels. The content provider may access the content provider interface 192 over the network 130. For example, the content provider may use a web browser that runs on the content provider system 142. By accessing the application development tool using the content provider interface 192, the content provider may create interaction sites and interaction pages that will be used by the call handling system 150 when processing a request to the interaction site being created by the content provider. In the context of this discussion, a “page” is a discrete programming routine configured to perform a discrete function. A page may be defined by a user through an interaction with, for example, a GUI in which the user may 10 indicate the type of programming routine for the page and may optionally further indicate one or more other pages linked to the page. Processing may then proceed to the one or more other linked pages after completion of execution of the page or, alternatively, after initiation of execution of the page but before completion of execution of the page. A page may be compiled into one or more programming language 15 modules or scripts after the page is defined by the user through interaction with the
GUI. The one or more programming language modules or scripts may be used, for example, by a handling system to execute the discrete programming routine to thereby perform the discrete function of the page. Examples of different pages include message pages, question pages, logic pages, transaction pages, and multimodal 20 action pages. These different pages are described in further detail in issued US Patent
No. 8,582,727, which is incorporated herein by reference for all purposes.
An interaction page is a particular type of page that is configured to perform the function of delivering content to and/or receiving content from a user via a communications channel used by the user to contact a multi-channel system (e.g., 25 voice communications channel for telephone contacts, chat communications channel for chat contacts, email communications channel for email contacts, and SMS communications channel for SMS contacts). A “voice page” is a particular type of interaction page that is configured to perform the function of delivering audible content
-192016246072 28 Mar 2019 to and/or receiving audible content from a user that called a telephone number assigned to the voice site. A VIVR-enabled voice page may further provide and/or receive visual content from the user that accesses the voice site using the communications device 110. FIGS. 2A-2B, 3A-3B, and 4A-4B, 5A-5B, and 6A-6B 5 illustrate examples of one or more pages provided by a GUI of an application development tool.
The interaction sites and pages created by the content provider using the content provider interface 192 are interpreted and/or compiled by an application compiler 194 to generate scripts that are executed by the multi-channel communications system 10 interacting with a user accessing the interaction site. In some implementations, the application compiler 194 may generate an interaction flow document, which may include XML scripts or code that correspond to pages (i.e., programming modules) of an interaction site created by the content provider. For example, the XML scripts may be SCXML scripts. The interaction flow document may be stored in an interaction flow 15 document database 185. The interaction flow processor 180 may access the scripts from the interaction flow document database 185 and translate them into a language that can be processed by a particular handling system when the call handling system 150 interacts with a user accessing a voice site.
In addition to the XML scripts, the application compiler 194 may also generate other types of scripts (e.g. Java scripts) and other types of executable code using other programming languages based on pages created for the interaction site by the content provider (e.g., based on transaction pages). The other types of scripts may be used by the call handling system 150 to interact over the data network 130 with the user accessing the interaction site.
The data store 160 is configured to store user interaction data with interaction sites. In some implementations, the data store 160 may store interaction data associated with a particular user. For example, the interaction data may include the gender and other characteristics of the user, the choices made by the user during each state of the
-202016246072 28 Mar 2019 interaction, and the resources utilized during each state of the interaction. In some implementations, the data store 160 may store aggregated interaction data associated with a particular interaction site. For example, the aggregated interaction data may include data specifying a breakdown of genders among all users that accessed the 5 particular interaction site. In some implementations, a user may opt-out such that her usage data is then not stored in the data store 160. In some implementations, a user may opt-in to have her usage data be stored in the data store 160.
FIGS. 2A-2B illustrate a GUI 200 for an application development tool that is used by a content provider to create a VIVR-enabled interaction site. In general, each interaction 10 site includes a flow of the interaction states that provide an overview of how users interact with the interaction site during the execution of the interaction site. A state may be configured using a page, such as, for example, a voice page or, more generally, an interaction page. In some implementations, the states of the flow for an interaction site are the same across multiple communications channels. For example, 15 a first user may access a VIVR-enabled voice site by calling a telephone number associated with the voice site, and in the first state, the first user would experience a “Say Greeting” voice page which greets the first user via voice. A second user may access the same voice site visually using an application by entering a URL associated with the voice site, and according to the same flow, the second user would also interact 20 with the “Say Greeting” interaction page which greets the second user via a visual interface. It may be a tedious process if the content provider is required to configure the same greeting message for both voice and visual interactions. The content provider interface 192 of the application builder 190 provides the content provider with a unified interface to create and configure pages that are common to both voice and 25 visual interactions without the need to enter duplicate information for these interactions.
The GUI 200 may be implemented by the content provider web interface 192 and presented to the content provider 142 when the content provider 142 accesses the application builder 190 using a web browser over the network 130 to create/manage
-212016246072 28 Mar 2019 the VIVR-enabled interaction site. The following describes the different components of the GUI 200 with respect to the system 100 that is described with reference to FIG. 1. Specifically, the components of the GUI 200 are described as used by the content provider 142 to create an interaction site for providing an ordering service to users of a product associated with the content provider 142. However, the GUI 200 and the associated application development tool may be used by other systems, content providers or application developers to create any interaction site to perform any desired automated interaction flow in response to a customer contact.
FIG. 2A illustrates an example GUI of an interaction site overview page 201 that is 10 presented to the content provider when the content provider logs into the application builder 190 (e.g., by inputting a user identifier and a password) to create an interaction site using the content provider system 142. The Site Overview page 201 may be accessed by the content provider, for example, by selecting an interaction site from among a displayed list of interaction sites associated with (e.g., designed by or for) the 15 content provider, clicking on the desired interaction site (e.g., test app) and then clicking on the “Site Overview” link 201a. The Site Overview page 201 provides a listing of the different pages created by the content provider to define the interaction site. The Site Overview page 201 lists all the pages that are included in the interaction site. The name of the interaction site is specified in the heading 201b of the Site 20 Overview page 201 (e.g., test app).
When the user of communications device 110 interacts with the interaction site, the first page that is processed is identified in the “Home Page” field 201c. The content provider may specify any page that the content provider wants to be processed first as the Home Page 201c. In some implementations, the first page in the listing of pages is 25 the same page that is listed as the “Home Page” 201c. However, in other implementations, the page that is as the “Home Page” 201c is not the first page in the listing of the pages in the Site Overview page 201.
-222016246072 28 Mar 2019
The order in which the various pages are processed is determined by the links in the respective pages. Each page usually contains a link to the next page that is to be processed. For example, the interaction site illustrated in the Site Overview page 201 has a page flow 202 of ten interaction pages, including the interaction pages “Say
Greeting”, “Account”, “Account Number”, “Set up”, “Main Select”, “Billing”, “Sales”, “Support”, “Transfer”, and “Goodbye.” Each of the pages may be identified by a page name that is shown in the Page Name field 201 d. In addition or as an alternative to the page name, each page also may be identified by a page number that is shown in the Page # field 201 e. The page name and page number of a page are specified by the 10 content provider when creating the pages for the interaction site. A page may have a unique page name, or it may have a page name that is similar to the page name of another page. In case two or more pages share the same page name, they may be differentiated based on the page numbers. The combination of page name and page number uniquely identifies a page. A user may access and modify any of the pages 15 displayed in the page flow 202 by selecting them from the displayed list. In addition, each page may be associated with a page type as shown in the Page Type field 201k. The types of pages that may be created by the content provider using the application development tool may include, for example: (1) a message page; (2) a question page;
(3) a logic page; (4) a transaction page; (5) a data page; and (6) a multimodal action 20 page. These different pages are described in further detail in issued US Patent No.
8,582,727, which is incorporated herein by reference for all purposes.
Importantly, a second page that is processed after the processing of a first page may be said to be directly linked to that first page if the first page includes a direct link to the second page without any intervening pages therebetween. Alternatively, a second 25 page may instead be said to be indirectly linked to the first page if that first page is linked to the second page in the interaction flow with one or more pages being processed between the processing of the first page and the processing of the second page.
-232016246072 28 Mar 2019
In other implementations, the Site Overview page 201 may additionally or alternatively present the pages in a two dimensional or three dimensional display that visually depicts the links between the pages. For example, each page may be displayed as a page graphical element, such as, for example, a rectangle or a block, with one or more 5 link graphical elements, such as, for example, lines, connecting the page graphical elements to other page graphical elements to which they are linked. Text may be overlaid on or displayed in proximity to the page and/or line graphical elements to communicate the identity of the corresponding page and/or the nature of the link between the elements.
The content provider may create a new page by clicking the “Add Page” button icon 201f. When the “Add Page” button icon 201f is clicked, a new page is added to the page flow 202. In response to selecting the button icon 201f, the GUI 200 may present a set of page templates for selection in, for example, a drop-down list. The page templates may include, for example, message pages, question pages, logic pages, transaction pages, and multimodal action pages. The user may select a page template from the list to generate a page of the corresponding type using the template. The template presents to the user the necessary fields and/or controls for that page type and the user may populate the fields (e.g., by typing text into the fields) and/or select the controls to generate a page of the corresponding type.
Alternatively, a new page may be created by copying a previously created page. The content provider may select the page to be copied by checking the checkbox to the left of the page to be copied and then selecting the “Copy” button. An existing page can be deleted by checking the checkbox to the left of the page, and then clicking the “Delete” button. The content provider may save the interaction site by clicking the 25 “Save” button 201j. The content provider may save the interaction site and then exit the GUI 201 by clicking the “Save & Exit” button 201g. The content provider may exit the GUI 201 without saving the interaction site by clicking the “Exit” button 201 h. The content provider may compile the interaction site by clicking the “Compile” button 201 i,
-242016246072 28 Mar 2019 where the application compiler 194 may compile the input parameters into an interaction flow document, and may save the interaction flow document in the interaction flow document database 185.
FIG. 2B illustrates an example GUI 204 for defining the contact information for an interaction site. For example, the content provider may click on the “Assign Contact” link 203 to access the Contact page 204. The Contact page 204 is specified by the heading 204a, and the corresponding interaction site is specified by the Site Name field 204b. The Contact page 204 allows the content provider to enable voice IVR and/or visual IVR. Although not shown, additionally or alternatively, the Contact page 10 204 may also provide a listing of communications channels (e.g., SMS, IVR, Chat, and/or Email) that users may use to access the interaction site. The content provider may enable the type of IVR the users may use to access the interaction site by checking or unchecking the checkbox 204c associated with the IVR type. The content provider may select one or more IVR type that the interaction site will support, and, for 15 each of the selected IVR type, the content provider may enter corresponding contact information to associate the contact information with the interaction site.
For the Voice IVR 206, the content provider may enter a Phone Number 206a (e.g., “1800-123-4567”), and click the Add New Phone Number button 206b to associate the entered phone number with the interaction site. When the user of the communications 20 device 110 enters the phone number on the communications device 110, the communications device 110 would communicate with the call handling system 150 and the interaction site would be launched, facilitating a voice interaction between the VIVR system 170 and the communications device 110.
For the Visual IVR 207, the content provider may enter a URL 207a (e.g., 25 “www.example.com/vivr/1 -800-123-4567”), and click the Add New VIVR Application button 207b to associate the entered URL with the interaction site. When the user of the communications device 110 enters the URL on the communications device 110, the communications device 110 would communicate with the call handling system 150
-252016246072 28 Mar 2019 and the interaction site would be launched, facilitating a visual interaction between the VIVR system 170 and the communications device 110.
In the examples above, more types of communications channels may be added to the Contact page 204 depending on the requirements of the interaction sites and the capability of the multi-channel communications system 135. One or more additional parameters may be added or substituted for any of the channels in the examples above. For example, there may be multiple phone numbers (e.g., toll-free and local phone numbers) associated with the voice IVR 206.
The “Connect to” drop-down menu 212 allows the content provider to choose which of the pages in the page flow 202 to be processed when the user of communications device 110 accesses the interaction site using the information in the Contact page. In some implementations, the “Connect to” drop-down menu 212 defines the “Home Page” field 201c in FIG 2A. Here, the content provider chooses the “Say Greeting” interaction page 10 as the home page of the interaction site. In some implementations, the user may select a different connect to page for different contacts (e.g., an email contact is connected to one page while a chat contact is connected to a different page).
The GUI 200 further includes a display 209 that illustrates all, one or a subset (e.g., all telephone contact points or all URL contact points) of the contact points associated 20 with the interaction site. In some implementations, the display 209 shows a list of all of the contact points assigned to the interaction site by the user through interactions with GUI 204. The display 209 may be dynamically updated in response to and as the user interacts with the GUI 204 to add or remove contact points from the display. In the example shown in Fig. 2B, two different contact points have been assigned to the 25 interaction site: (1) A telephone contact point, which is the telephone number 1-800123-4567; and (2) a VIVR contact point, which is the URL “www.example.com/vivr/1800-123-4567.”
-262016246072 28 Mar 2019
FIG. 2C illustrates an example GUI 220 for defining the site properties for an interaction site. For example, the content provider may click on the “Site Properties” link 221 to access the Site Properties page 220. The Site Properties page 220is specified by the heading 222. The Site Properties page 220 allows the content provider to input values for one or more parameters for the specified interaction site. For example, the VIVR Timeout parameter 223 allows the content provider to specify in the field 223a an amount of idle time for a VIVR session to expire. As another example, the Help URL parameter 224 allows the content provider to specify in the field 224a a URL that users can click on during a VIVR session to reach a help page (e.g., a Commonly Asked Question website, etc.). As another example, the Contact Us parameter 225 allows the content provider to specify in the field 225a a phone number that users can click on during a VIVR session to be directed to a voice interaction with a human agent or another voice application. As another example, the Logo parameter 226 allows the content provider to click on the Upload button 226a to upload a logo of the content provider to be displayed on the communications device 110 during a VIVR session. As another example, the Theme parameter 227 allows the content provider to click on the Upload button 227a to upload a theme (e.g., text color, background color and images, etc.) to be displayed during a VIVR session.
FIG. 3A illustrates an example GUI 301 for configuring an interaction page (specifically, a message page) that is the first page that is processed for the interaction site in the event that the user accesses the interaction site via an IVR channel. The “Say Greeting” interaction page is identified by its page name 302 and/or page number 303. The page name 302 and the page number 303 correspond to the name of the page shown in the Page Name field 201 d and the number of the page shown in the Page # field 201 e respectively, shown in the Site Overview page 201 in Fig. 2A. Here, the GUI
301 allows the content provider to input parameters that configure the interaction page for the IVR channel, including both voice and visual interactions. As described below,
-272016246072 28 Mar 2019 some of the input parameters are common for both voice and visual interactions, while some of the input parameters are specific for either the voice or the visual interaction.
The main tab 304a of the interaction page allows the content provider to enter parameters that configure the interaction page for the IVR channel. In some 5 implementations, some of these parameters in the GUI 301 are common for both voice and visual interactions. For example, the “Initial Prompts” parameter 304b allows the content provider to enter information that will be presented to the user independent of how the user accesses the VIVR-enabled voice site. The content provider inputs a prompt “Welcome to this test app!” in the “Initial Prompts” parameter 304c, where this 10 prompt will be played to the user as an audio message if the interaction is voice, or displayed as a text message on the communications device 110 if the interaction is visual.
As another example, the next page parameter 304d is also common for both voice and visual interactions. The next page parameter 304d includes a pull-down menu that 15 allows the content provider to specify which page is to be provided to the user after the initial prompts message is delivered to the user. In some implementations, the pages included in the pull-down menu reference to the pages in the page flow 202 defined by the content provider in Fig. 2A.
In some implementations, some of these parameters in the GUI 301 are specific for the 20 IVR channel. For example, if the content provider clicks on the “Text-to-Speech” button 304e, another GUI (not shown) may be provided to the content provider, which enables the content provider to configure parameters that are related to the automated speech (e.g., gender, pitch, speed, etc.) converted from the text specified in the “Initial Prompts” parameter 304b.
The “No Output” tab 305 of the interaction page allows the content provider to enter parameters that configure the interaction page for the IVR channel in the case where there is no response received from the user. Similar to the main tab 304a, some parameters in the “No Output” tab 305 are common for both voice and visual
-282016246072 28 Mar 2019 interactions, and some parameters in the “No Output” tab 305 are specific for the voice or the visual interaction. The “No Match” tab 306 of the interaction page allows the content provider to enter parameters that configure the interaction page for the IVR channel in the case where the received response does not match a specified response 5 type. Similar to the main tab 304a, some parameters in the “No Match” tab 306 are common for both voice and visual interactions, and some parameters in the “No Match” tab 306 are specific for the voice or the visual interaction.
The examples describe above for the IVR channel are not limiting. A GUI for developing an interaction page used in an IVR communications environment may have 10 more or fewer configurable parameters than the GUI 301 described in FIG. 3A.
FIG. 3B illustrates an example GUI 311 for configuring parameters associated with a visual interaction of the first page that is processed for the interaction site in the event that the user initiates a visual interaction with the interaction site. The Visual Settings tab 314a of the interaction page allows the content provider to enter values for 15 parameters associated the visual interaction. These parameters will not be processed if the user access the interaction site through a voice interaction. For example, the “Title” parameter 314b allows the content provider to enter information that will be presented to the user as a title of the voice page when the user accesses the VIVRenabled voice site through a visual interaction. The content provider inputs text 20 “Welcome to Example Store!” in the field 314c, where this text will be displayed on the communications device 110 if the interaction is visual.
As another example, the “Navigations Controls” parameter 314d allows the content provider to specify navigation controls through check boxes to allow the user to navigate the voice pages according to the interaction flow of the interaction site. Here, 25 the content provider selects the “Next” checkbox, where a “Next” button will be displayed on the communications device 110 if the interaction is visual to allow the user to navigate to the next voice page in the interaction flow.
-292016246072 28 Mar 2019
As another example, the “Other Controls” parameter 314f allows the content provider to specify additional controls for the interaction site during a visual interaction. The “Call” checkbox 314g allows a “Call” button to be displayed on the communications device 110 if the interaction is visual. If the user clicks on the “Call” button during a 5 visual interaction with the voice page, the communications device 110 may dial the telephone number as specified in the “Contact Us” parameter 225 in Fig. 2C to connect the user to a human agent or another voice site. The “Help” checkbox 314h allows a “Help” button to be displayed on the communications device 110 if the interaction is visual. If the user clicks on the “Help” button during a visual interaction with the voice 10 page, the communications device 110 may open a web site as specified in the “Help
URL” parameter 224 in Fig. 2C to connect the user to a help web site. The examples describe above for configuring the visual interaction are not limiting. A GUI for developing an interaction page used in a VIVR communications environment may have more or fewer configurable parameters than the GUI 311 described in FIG. 3B.
FIG. 3C illustrates an example GUI 321 that is displayed on a communications device
110 during a visual interaction with the first page of the VIVR-enabled voice site. In some implementations, the example GUI 321 may be displayed using a web browser running on the communications device 110. In some implementations, the example GUI 321 may be displayed using a third-party mobile application running on the 20 communications device 110. In some implementations, the arrangement of the displayed information may be defined by scripts or codes received from the call handling system 150. For example, the arrangement of the displayed information may be defined using HTML5 codes transmitted from the application server 140 to the communications device 110, which allows flexibility for how the information is being 25 displayed according to parameters such as device types, device screen size, user profiles, and/or other variable parameters.
In some implementations, the example GUI 321 may display information that is common as the audio information provided to a user during a voice interaction with the
-302016246072 28 Mar 2019 first page of the VIVR-enabled voice site. For example, the main text 325 “Welcome to this test app!” is displayed on the communications device 110, which corresponds to the information specified by the content provider in the “Initial Prompts” field 304c in Fig. 3A.
In some implementations, the example GUI 321 may display information that is only applicable for a visual interaction with the first page of the VIVR-enabled voice site. For example, a logo 322 is displayed on the communications device 110, which corresponds to the “Logo” parameter 226 as illustrated in Fig. 2C. As another example, a “Next” button 323 is displayed on the communications device 110, which 10 corresponds to the selection of the “Next” parameter 314e as illustrated in Fig. 3B. As another example, a title text 324 of “Welcome to Example Store” is displayed on the communications device 110, which corresponds to the text entered in the “Title” parameter 314c as illustrated in Fig. 3B. As another example, a “Help” button 326 is displayed on the communications device 110, which corresponds to the selection of the 15 “Help” parameter 314h as illustrated in Fig. 3B. As another example, a “Contact Us” button 327 is displayed on the communications device 110, which corresponds to the selection of the “Call” parameter 314g as illustrated in Fig. 3B.
FIG. 4A illustrates an example GUI 401 for configuring an interaction page (specifically, a question page) that is the second page that is processed for the interaction site. The 20 “Account” interaction page is identified by its page name 402 and/or page number 403.
The main tab 404a of the interaction page allows the content provider to enter parameters that configure the interaction page for the IVR channel. The question that is asked is specified by the content provider using the “Initial Prompts” parameter 404b. The response received from the caller is processed based on the “Response Type” 25 405a specified by the content provider and is stored in a variable 405b.
In some implementations, some of these parameters in the GUI 401 are common for both voice and visual interactions. For example, the “Initial Prompts” parameter 404b allows the content provider to enter information that will be presented to the user
-312016246072 28 Mar 2019 independent of how the user accesses the VIVR-enabled voice site. As another example, the Response Type 405a is also common for both voice and visual interactions. Here, the Response Type 405a is set to “Yes/No”, where the call handling system 150 expects a Yes or No response from a user accessing this voice page independent of whether the interaction is visual or voice. As another example, the next page parameter 404d is also common for both voice and visual interactions. The next page parameter 404d includes two pull-down menus that allows the content provider to specify which page is to be provided to the user after the user provides a feedback.
The examples describe above for the IVR channel are not limiting. A GUI for 10 developing an interaction page used in an IVR communications environment may have more or fewer configurable parameters than the GUI 401 described in FIG. 4A.
FIG. 4B illustrates an example GUI 411 for configuring parameters associated with a visual interaction of the second page that is processed for the interaction site in the event that the user initiates a visual interaction with the interaction site. The Visual 15 Settings tab 414a of the interaction page allows the content provider to enter values for parameters associated the visual interaction. These parameters will not be processed if the user access the interaction site through a voice interaction. The “Title”, “Navigations Controls”, and “Other Controls” parameters are similar to those described in Fig. 3B and will not be repeated here. The “VIVR Display Option” 415 is shown in 20 the GUI 411 in response to the content provider selecting the Response Type 405a to “Yes/No.” The “VIVR Display Option” 415 allows the content provider to input additional text in fields 415a and 415b to be displayed to a user to further clarify the Yes/No choices during a visual interaction.
The examples describe above for configuring the visual interaction are not limiting. A 25 GUI for developing an interaction page used in a VIVR communications environment may have more or fewer configurable parameters than the GUI 411 described in FIG. 4B.
-322016246072 28 Mar 2019
FIG. 4C illustrates an example GUI 421 that is displayed on a communications device 110 during a visual interaction with the second page of the VIVR-enabled voice site. In some implementations, the arrangement of the displayed information may be defined by scripts or codes received from the call handling system 150. For example, the 5 arrangement of the displayed information may be defined using HTML5 codes transmitted from the application server 140 to the communications device 110. In some implementations, the arrangement of the displayed information is defined by parameters such as the page type, the response type, and the Visual Settings options selected by the content provider. For example, the arrangement of the displayed 10 information in GUI 421 is defined by the page type being a question page, the response type being a “Yes/No” type, and the selection of parameters in the Visual Settings tab 414a in Fig. 4B.
In some implementations, the example GUI 421 may display information that is common as the audio information provided to a user during a voice interaction with the 15 corresponding voice page of the VIVR-enabled voice site. In some implementations, the example GUI 421 may display information that is only applicable for a visual interaction with the corresponding voice page of the VIVR-enabled voice site. For example, a logo 422 is displayed on the communications device 110, which corresponds to the “Logo” parameter 226 as illustrated in Fig. 2C. As another 20 example, a “Next” button 429, a “Back button” 423, a title text 424, a “Help” button 426, a “Contact Us” button 427 are displayed on the communications device 110, which corresponds to the selections as illustrated in Fig. 4B. As another example, a text 428a of “Yes, I do” is displayed on the communications device 110, which corresponds to the text entered in the “Yes” parameter 415a as illustrated in Fig. 4B. As another 25 example, a text 428b of “No, I do not” is displayed on the communications device 110, which corresponds to the text entered in the “No” parameter 415b as illustrated in Fig. 4B. After the user makes a selection, the user can click on the “Next” button 429, where a value corresponding to the selection is stored to a corresponding variable,
-332016246072 28 Mar 2019 transmitted to the call handling system 150, and processed by the VIVR system 152 in a same manner that a response from a voice interaction would be processed.
FIG. 5A illustrates an example GUI 501 for configuring an interaction page (specifically, a question page) that corresponds to page number 30 of the interaction site. The 5 “Account Number” interaction page is identified by its page name 502 and/or page number 503. In some implementations, some of these parameters in the GUI 501 are common for both voice and visual interactions. For example, the “Initial Prompts” parameter 504b allows the content provider to enter information that will be presented to the user independent of how the user accesses the VIVR-enabled voice site. As 10 another example, the Response Type 505a is also common for both voice and visual interactions. Here, the Response Type 505a is set to “Number”, where the call handling system 150 expects a numeric response from a user accessing this voice page independent of whether the interaction is visual or voice. As another example, the next page parameter 504d is also common for both voice and visual interactions.
The examples describe above for the IVR channel are not limiting. A GUI for developing an interaction page used in an IVR communications environment may have more or fewer configurable parameters than the GUI 501 described in FIG. 5A.
FIG. 5B illustrates an example GUI 511 for configuring parameters associated with a visual interaction of interaction page that corresponds to page number 30 of the 20 interaction site in the event that the user initiates a visual interaction with the interaction site. The Visual Settings tab 514a of the interaction page allows the content provider to enter values for parameters associated the visual interaction. These parameters will not be processed if the user access the interaction site through a voice interaction. The “Title”, “Navigations Controls”, and “Other Controls” parameters are similar to 25 those described in Fig. 3B and will not be repeated here. Unlike Fig. 4B, the GUI 511 does not include the “VIVR Display Option” parameter because the content provider selecting the Response Type 505a to “Number.” The examples describe above for configuring the visual interaction are not limiting. A GUI for developing an interaction
-342016246072 28 Mar 2019 page used in a VIVR communications environment may have more or fewer configurable parameters than the GUI 511 described in FIG. 5B.
FIG. 5C illustrates an example GUI 521 that is displayed on a communications device 110 during a visual interaction with the voice page number 30 of the VIVR-enabled 5 voice site. In some implementations, the arrangement of the displayed information may be defined by scripts or codes received from the call handling system 150. For example, the arrangement of the displayed information may be defined using HTML5 codes transmitted from the application server 140 to the communications device 110. In some implementations, the arrangement of the displayed information is defined by 10 parameters such as the page type, the response type, and the Visual Settings options selected by the content provider. For example, the arrangement of the displayed information in GUI 521 is defined by the page type being a question page, the response type being a “Number” type, and the selection of parameters in the Visual Settings tab 514a in Fig. 5B.
In some implementations, the example GUI 521 may display information that is common as the audio information provided to a user during a voice interaction with the corresponding voice page of the VIVR-enabled voice site. In some implementations, the example GUI 521 may display information that is only applicable for a visual interaction with the corresponding voice page of the VIVR-enabled voice site. For example, a logo 522 is displayed on the communications device 110, which corresponds to the “Logo” parameter 226 as illustrated in Fig. 2C. As another example, a “Next” button 529, a “Back button” 523, a title text 524, a “Help” button 526, a “Contact Us” button 527 are displayed on the communications device 110, which corresponds to the selections as illustrated in Fig. 5B. As another example, a text field
528 is displayed on the communications device 110, which allows the user to input numerical values that corresponds to an account number. After the user is finished entering the account number, the user can click on the “Next” button 529, where the input numerical values are stored in a corresponding variable, transmitted to the call
-352016246072 28 Mar 2019 handling system 150, and processed by the VIVR system 152 in a same manner that a response from a voice interaction would be processed.
FIG. 6A illustrates an example GUI 601 for configuring an interaction page (specifically, a question page) that corresponds to page number 50 of the interaction site. The “Main Select” interaction page is identified by its page name 602 and/or page number
603. In some implementations, some of these parameters in the GUI 601 are common for both voice and visual interactions. For example, the “Initial Prompts” parameter 604b allows the content provider to enter information that will be presented to the user independent of how the user accesses the VIVR-enabled voice site. As another 10 example, the Response Type 605a is also common for both voice and visual interactions. Here, the Response Type 605a is set to “Keyword”, where the call handling system 150 expects a keyword from one or more predefined keywords 631a, 631b, and 631c, independent of whether the interaction is visual or voice. As another example, the next page parameter 604d that corresponds to a selection by the user is 15 also common for both voice and visual interactions.
The examples describe above for the IVR channel are not limiting. A GUI for developing an interaction page used in an IVR communications environment may have more or fewer configurable parameters than the GUI 601 described in FIG. 6A.
FIG. 6B illustrates an example GUI 611 for configuring parameters associated with a visual interaction of interaction page that corresponds to page number 50 of the interaction site in the event that the user initiates a visual interaction with the interaction site. The Visual Settings tab 614a of the interaction page allows the content provider to enter values for parameters associated the visual interaction. These parameters will not be processed if the user access the interaction site through a voice interaction.
The “Title”, “Navigations Controls”, and “Other Controls” parameters are similar to those described in Fig. 3B and will not be repeated here. The “VIVR Display Options”
615 is shown in the GUI 611 in response to the content provider selecting the Response Type 605a to “Keyword.” The “VIVR Display Option” 615 allows the content
-362016246072 28 Mar 2019 provider to select from a menu of options that controls the manners the keywords are to be displayed to a user during a visual interaction. Example options may include “Click to continue,” “Checkboxes,” and “Radio buttons.” Here, the option “Click to continue” is chosen by the content provider, which would allow the user to select one keyword from the list of predetermined keywords 631a, 631b, and 631c, as defined in
Fig. 6A. The examples describe above for configuring the visual interaction are not limiting. A GUI for developing an interaction page used in a VIVR communications environment may have more or fewer configurable parameters than the GUI 611 described in FIG. 6B.
FIG. 6C illustrates an example GUI 621 that is displayed on a communications device
110 during a visual interaction with the voice page number 50 of the VIVR-enabled voice site. In some implementations, the arrangement of the displayed information may be defined by scripts or codes received from the call handling system 150. For example, the arrangement of the displayed information may be defined using HTML5 15 codes transmitted from the application server 140 to the communications device 110.
In some implementations, the arrangement of the displayed information is defined by parameters such as the page type, the response type, and the Visual Settings options selected by the content provider. For example, the arrangement of the displayed information in GUI 621 is defined by the page type being a question page, the 20 response type being a “Keyword” type, and the selection of parameters in the Visual
Settings tab 614a in Fig. 6B.
In some implementations, the example GUI 621 may display information that is common as the audio information provided to a user during a voice interaction with the corresponding voice page of the VIVR-enabled voice site. In some implementations, 25 the example GUI 621 may display information that is only applicable for a visual interaction with the corresponding voice page of the VIVR-enabled voice site. For example, a logo 622 is displayed on the communications device 110, which corresponds to the “Logo” parameter 226 as illustrated in Fig. 2C. As another
-372016246072 28 Mar 2019 example, a “Next” button 629, a “Back button” 623, a title text 624, a “Help” button 626, a “Contact Us” button 627 are displayed on the communications device 110, which corresponds to the selections as illustrated in Fig. 6B. As another example, a list of keyword selections 628a, 628b, and 628c are displayed on the communications device 5 110, which allows the user to select one keyword from the list of keyword selections
628a, 628b, and 628c. After the user is finished selecting the keyword, the user can click on the “Next” button 629, where the value corresponding to the selected keyword is stored in a corresponding variable, transmitted to the call handling system 150, and processed by the VIVR system 152 in a same manner that a response from a voice 10 interaction would be processed.
Figs. 7-10 show additional example question voice pages with various response types that are rendered and displayed on a communications device, to illustrate the design flexibility of VIVR applications. Fig. 7 shows a GUI 701 that is displayed on a communications device 110 during a visual interaction with a question voice page 15 assigned with a response type of “multiple selections.” A list of keyword selections
703a, 703b, 703c, and 703d are displayed on the communications device 110, which allows the user to select one or more keywords from the list of keyword selections 703a, 703b, 703c, and 703d by clicking on the corresponding checkboxes. After the user is finished selecting the keyword, the user can click on the “Next” button 729, 20 where the value(s) corresponding to the selected keyword(s) is stored in one or more corresponding variables, transmitted to the call handling system 150, and processed by the VIVR system 152 in a same manner that a response from a voice interaction would be processed.
Fig. 8 shows a GUI 801 that is displayed on a communications device 110 during a visual interaction with a question voice page assigned with a response type of “location.” A text field 803a and a menu 803b are displayed on the communications device 110, which allows the user to input location information. After the user is finished inputting the location information, the user can click on the “Next” button 829,
-382016246072 28 Mar 2019 where the value corresponding to the input location information is stored in a corresponding variable, transmitted to the call handling system 150, and processed by the VIVR system 152 in a same manner that a response from a voice interaction would be processed.
Fig. 9 shows a GUI 901 that is displayed on a communications device 110 during a visual interaction with a question voice page assigned with a response type of “calendar.” A day field 903a and month and year menus 903b and 903c are displayed on the communications device 110, which allows the user to input calendar information. After the user is finished inputting the calendar information, the user can 10 click on the “Next” button 929, where the value corresponding to the input calendar information is stored in a corresponding variable, transmitted to the call handling system 150, and processed by the VIVR system 152 in a same manner that a response from a voice interaction would be processed.
Fig. 10 shows a GUI 1001 that is displayed on a communications device 110 during a visual interaction with a question voice page assigned with a response type of “calendar.” A “Card Number” field 1003a and “Expiration Month” and “Expiration Year” menus 1003b and 1003c are displayed on the communications device 110, which allows the user to input payment information. After the user is finished inputting the calendar information, the user can click on the “Next” button 1029, where the value 20 corresponding to the input payment information is stored in a corresponding variable, transmitted to the call handling system 150, and processed by the VIVR system 152 in a same manner that a response from a voice interaction would be processed.
FIG. 11 is a flow chart illustrating an example of a process 1100 for a content provider to create and host a VIVR-enabled interaction site. In general, the process 1100 25 provides a content provider an interface to design and create an interaction site that may be accessed in a multi-channel solution platform. The process 1100 is described as being performed by a computer system comprising one or more computers, for example, the content provider system 142 shown in FIG. 1.
-392016246072 28 Mar 2019
The application builder 190 provides to the content provider system 142 a content provider interface for developing an interaction site executable by a user device to communicate with a server using any one of multiple communications channels (1101).
The content provider system 142 creates a VIVR-enabled interaction site using the content provider interface (1103). The application builder 190 may include a content provider interface 192. The content provider interface 192 is a GUI front-end for an application development tool that can be used to build an interaction site that is capable of handling interactions using multiple communications channels. The content provider may access the content provider interface 192 over the network 130. For example, the application builder 190 may provide a GUI similar to the GUI 204 in FIG. 2B, which allows the content provider to enable voice IVR and/or visual IVR for an interaction site.
The content provider system 142 inputs contact data for each of the enabled
IVRs(1105). For example, the application builder 190 may provide a GUI similar to the
GUI 204 in FIG. 2B, which allows the content provider to define the contact information for an interaction site. The content provider may enable the type of IVR the users may use to access the interaction site by checking or unchecking the checkbox associated with the IVR type. The content provider may select one or more IVR type that the interaction site will support, and, for each of the selected IVR type, the content provider may enter corresponding contact information to associate the contact information with the interaction site. For the Voice IVR, the content provider may enter a Phone Number. For the Visual IVR, the content provider may enter a URL.
The content provider system 142 determines whether the content provider has finished defining pages of the interaction site (1107). For example, the application builder 190 may provide a GUI similar to the GUI 201 in FIG. 2A, which allows the content provider to add, edit, or remove one or more pages of an interaction site. The pages may include, for example, any or all of the previously mentioned pages, including message pages, question pages, logic pages, and transaction pages. Examples of other page
-402016246072 28 Mar 2019 types that may be used for the interaction site include call queue pages, which are configured to interact with external data sources in order to pull or push relevant data and call transfer pages, which are configured to transfer the call to designated contact point (e.g., phone number). If the content provider has not finished defining the pages 5 of the interaction site, the content provider selects a page from among the site's existing pages to edit (e.g., by modifying the page or deleting the page) or generates a new page(1109).
The content provider inputs page data for the selected or new page (1111). For example, the application builder 190 may provide a GUI similar to the GUI 301 in FIG.
3A for configuring an interaction page that is the first page processed for the interaction site for an IVR channel. Some of the input parameters are common for both voice and visual interactions, while some of the input parameters are specific for either the voice or the visual interaction. Optionally, the content provider inputs visual-interactionspecific data for the selected or new page (1113). For example, the application builder
190 may provide a GUI similar to the GUI 311 in FIG. 3B that are specific for visual interactions for the IVR channel.
If the content provider has finished defining the pages of the interaction site, the application builder 190 generates an interaction flow document including data defining the interaction site based on the contact data and the page data (1115). In some 20 implementations, the application builder 190 may generate an interaction flow document, which may include XML scripts that correspond to pages (e.g., interaction pages) of an interaction site created by the content provider via the content provider interface.
The application builder 190 then stores the interaction flow document at the interaction 25 flow document database 185 (1117). The interaction flow document database 185 stores interaction flow documents created by the application builder 190, and provides the interaction flow processor 180 access to these interaction flow documents.
-412016246072 28 Mar 2019
FIG. 12 is a flow chart illustrating an example of a process 1200 for a user to communicate with a call handling system and access an VIVR-enabled interaction site. In general, the process 1200 provides a communications device access to an interaction site to communicate with a VIVR system using either a voice interaction or a 5 visual interaction. The process 1200 is described as being performed by a computer system comprising one or more computers, for example, the communications system 100 shown in FIG. 1
The communications device 110 sends a request to access an interaction site using either a voice interaction or a visual interaction (1201). A user of the communications 10 device (e.g., a smart phone) 110 is able to interact with the communications device 110 to request a service from an interaction site that is provided by a content provider. For example, the user may indicate a desire to request a service by contacting the call handling system 150 in either of two ways. For example, a user may access a VIVRenabled voice site by calling a telephone number associated with the voice site. As 15 another example, the user may access the same voice site visually using an application by entering a URL associated with the voice site on the communications device.
The call handling system 150 receives the request from communications device 110, and sends a request to the interaction flow processor 180 to obtain code for initiating 20 the interaction site (1203). For example, if the user of the communications device 110 calls a phone number to reach the interaction site, the call handling system 150 will receive the phone call. Based on the contact information received by the call handling system 150, the call handling system 150 sends a request to the interaction flow processor 180 for the scripts for executing the voice site. The request sent by the call 25 handling system 150 to the interaction flow processor 180 may include an interaction site identifier (e.g., a unique interaction site identifier) that may be used by the interaction flow processor 180 to identify the desired interaction site. In some implementations, the call handling system 150 may send a request for the scripts for
-422016246072 28 Mar 2019 executing the entire flow of the interaction site. In other implementations, the call handling system 150 may send a request for the scripts for executing a particular state of the flow (e.g., a state corresponding to executing a single page or executing a subset of the pages of the interaction site), rather than the entire flow.
The interaction flow processor 180 identifies the communications channel used by the user to contact the system (1205). In some implementations, the communications channel may be included in the request sent by a multi-channel communications system. In some implementations, the communications channel may be determined by the interaction flow processor 180 based on the identifier of the handling system. For 10 example, the identifier may be an IP address of the handling system. As another example, the identifier may be metadata embedded in the request to the interaction flow processor 180.
The interaction flow processor 180 accesses the interaction flow document for the interaction site (1207). Based on the interaction site that the call handling system 150 15 has requested, the interaction flow processor 180 accesses the interaction flow document stored in the interaction flow document database 185. The interaction flow document database 185 then provides the common code for the interaction site (1209). In some implementations, the common code may be XML scripts. For example, the common code may be SCXML scripts.
The interaction flow processor 180 translates the common code to code specific to the IVR channel (1211). For example, given the handling system is the call handling system 150, the interaction flow processor 180 translates the scripts from XML scripts to VoiceXML scripts. In some implementations, the translation may include adding parameters specific to a type of communications channel in the translated scripts. For 25 example, given the handling system is the call handling system 150, the interaction flow processor 180 may add information specific to ASR resource selection in the translated scripts. The interaction flow processor 180 then transmits the translated code that is specific to the IVR channel to the call handling system 150 (1213).
-432016246072 28 Mar 2019
The call handling system 150 determines whether the request from the communications device 110 is for a standard IVR or for a visual IVR application (1214). In some implementations, the request from the communications system is received by the application server 140. For example, the application server 140 may determine 5 that the request is for a standard IVR if the request is received via a voice channel.
The application server 140 may determine that the request is for a visual IVR if the request is received via a data channel. As another example, the application server 140 may receive the request from a data channel, but may determine that the request is for a standard IVR if metadata associated with the request indicates that the request is a 10 voice communication. In some implementations, after the application server 140 determines whether the request is for a visual IVR or for a voice IVR, the application server 140 may send an updated request to the VIVR system 152, where the updated request includes an indication of the determination.
The call handling system 150 executes code specific to the determined type of IVR 15 communications to initiate the interaction site between the call handling system 150 and the communications device 110 (1215). The communications device 110 then interacts with the interaction site via the communications channel (1217). In some implementations, the VIVR system 152 processes a visual interaction in a same manner as processing a voice interaction. In some implementations, in response to 20 determining that the request from the communications device 110 is for the visualversion of the voice site, the application server 140 may translate the voice messages from the VIVR system 152 into corresponding messages in a different scripting or programming language (e.g., HTML5), so the corresponding messages may be displayed on the communications device 110. In some implementations, the call 25 handling system 150 may store interaction data in the data store 160, including data indicating whether the interaction is visual or voice.
The disclosed and other examples can be implemented as one or more computer program products, i.e., one or more modules of computer program instructions
-442016246072 28 Mar 2019 encoded on a computer readable medium for execution by, or to control the operation of, data processing apparatus. The implementations can include single or distributed processing of algorithms. The computer readable medium can be a machine-readable storage device, a machine-readable storage substrate, a memory device, or a 5 combination of one or more them. The term “data processing apparatus” encompasses all apparatus, devices, and machines for processing data, including by way of example a programmable processor, a computer, or multiple processors or computers. The apparatus can include, in addition to hardware, code that creates an execution environment for the computer program in question, e.g., code that 10 constitutes processor firmware, a protocol stack, a database management system, an operating system, or a combination of one or more of them.
A system may encompass all apparatus, devices, and machines for processing data, including by way of example a programmable processor, a computer, or multiple processors or computers. A system can include, in addition to hardware, code that 15 creates an execution environment for the computer program in question, e.g., code that constitutes processor firmware, a protocol stack, a database management system, an operating system, or a combination of one or more of them.
A computer program (also known as a program, software, software application, script, or code) can be written in any form of programming language, including compiled or 20 interpreted languages, and it can be deployed in any form, including as a standalone program or as a module, component, subroutine, or other unit suitable for use in a computing environment. A computer program does not necessarily correspond to a file in a file system. A program can be stored in a portion of a file that holds other programs or data (e.g., one or more scripts stored in a markup language document), in 25 a single file dedicated to the program in question, or in multiple coordinated files (e.g., files that store one or more modules, sub programs, or portions of code). A computer program can be deployed to be executed on one computer or on multiple computers
-452016246072 28 Mar 2019 that are located at one site or distributed across multiple sites and interconnected by a communications network.
The processes and logic flows described in this document can be performed by one or more programmable processors executing one or more computer programs to perform 5 functions by operating on input data and generating output. The processes and logic flows can also be performed by, and apparatus can also be implemented as, special purpose logic circuitry, e.g., an FPGA (field programmable gate array) or an ASIC (application specific integrated circuit).
Processors suitable for the execution of a computer program include, by way of 10 example, both general and special purpose microprocessors, and any one or more processors of any kind of digital computer. Generally, a processor will receive instructions and data from a read only memory or a random access memory or both. The essential elements of a computer can include a processor for performing instructions and one or more memory devices for storing instructions and data. 15 Generally, a computer can also include, or be operatively coupled to receive data from or transfer data to, or both, one or more mass storage devices for storing data, e.g., magnetic, magneto optical disks, or optical disks. However, a computer need not have such devices. Computer readable media suitable for storing computer program instructions and data can include all forms of nonvolatile memory, media and memory 20 devices, including by way of example semiconductor memory devices, e.g., EPROM,
EEPROM, and flash memory devices; magnetic disks, e.g., internal hard disks or removable disks; magneto optical disks; and CD ROM and DVD-ROM disks. The processor and the memory can be supplemented by, or incorporated in, special purpose logic circuitry.
While this document may describe many specifics, these should not be construed as limitations on the scope of an invention that is claimed or of what may be claimed, but rather as descriptions of features specific to particular embodiments. For example, a communications channel may include the Web, where a user may interact with an
-462016246072 28 Mar 2019 interaction site via a webpage generated dynamically according to the interaction flow. As another example, a communications channel may include a smart phone application, where a user may interact with an interaction site by starting a smart phone application, and the smart phone application then contacts the interaction site and provides a communications interface between the user and the interaction site. Certain features that are described in this document in the context of separate embodiments can also be implemented in combination in a single embodiment. Conversely, various features that are described in the context of a single embodiment can also be implemented in multiple embodiments separately or in any suitable sub-combination.
Moreover, although features may be described above as acting in certain combinations and even initially claimed as such, one or more features from a claimed combination can in some cases be excised from the combination, and the claimed combination may be directed to a sub-combination or a variation of a sub-combination. Similarly, while operations are depicted in the drawings in a particular order, this should not be understood as requiring that such operations be performed in the particular order shown or in sequential order, or that all illustrated operations be performed, to achieve desirable results.
Only a few examples and implementations are disclosed. Variations, modifications, and enhancements to the described examples and implementations and other 20 implementations can be made based on what is disclosed.

Claims (20)

1. A computer-implemented method, comprising:
receiving, by an application server and from a communications device, a request
2. The method of claim 1, wherein transmitting, from the application server and to the automated response system, the request to execute the interaction site comprises transmitting to the automated response system an indication that indicates whether the request from the communications device is for the voice-interaction with the interaction
3. The method of claim 1 or 2, wherein determining whether the request from the communications device is for the voice-interaction with the interaction site or the visualinteraction with the interaction site comprises:
20 determining whether the request is received from a voice channel or a data channel;
in response to determining that the request is received from the voice channel, determining that the request is for the voice-interaction with the interaction site; and in response to determining that the request is received from the data channel, 25 determining that the request is for the visual-interaction with the interaction site.
-492016246072 28 Mar 2019
4. The method of claim 1, 2 or 3, wherein determining whether the request from the communications device is for the voice-interaction with the interaction site or the visual-interaction with the interaction site comprises:
determining that the request is based on hypertext markup language (HTML);
5 receiving, from the automated response system, first instructions that, when executed, provide one or more steps of the multi-step communication flow between the communications device and the automated response system;
determining whether the request from the communications device is for a voiceinteraction with the interaction site or a visual-interaction with the interaction site;
5 determining whether the request from the communications device is for a voice-interaction with the interaction site or a visual-interaction with the interaction site;
in response to determining that the request from the communications device is for the voice-interaction with the interaction site:
retrieving the common interaction flow document;
5 based on state chart extensible markup language (SCXML), wherein the second instructions are based on voice extensible markup language (VoiceXML), and wherein the third instructions are based on HTML5.
5 and in response to determining that the request is based on HTML, determining that the request from the communications device is for the visual-interaction with the interaction site.
5 instructions , wherein the third instructions, when executed by the communications device, provide the one or more steps through a visual interaction with the user of the communications device; and, transmitting, from the application server and to the communications device, the third instructions.
5 to execute an interaction site for enabling communications to be exchanged between an automated response system and the communications device, wherein the interaction site specifies a multi-step communication flow between the communications device and the automated response system;
transmitting, from the application server and to the automated response system, 10 a request to execute the interaction site;
receiving, by the application server and from the automated response system, a common interaction flow document including first instructions that, when executed, provide one or more steps of the multi-step communication flow between the communications device and the automated response system, wherein the common 15 interaction flow document is for handling interactions received over different channels including voice interactions and visual interactions;
determining, by the application server, whether the request from the communications device is for a voice-interaction with the interaction site or a visualinteraction with the interaction site;
20 in response to determining that the request from the communications device is for the voice-interaction with the interaction site:
retrieving the common interaction flow document;
translating, by the application server, the first instructions into second instructions , wherein the second instructions, when executed by the communications 25 device, provide the one or more steps through a voice interaction with a user of the communications device; and transmitting, from the application server and to the communications device, the second instructions; and
-482016246072 28 Mar 2019 in response to determining that the request from the communications device is for the visual-interaction with the interaction site:
retrieving the common interaction flow document;
translating, by the application server, the first instructions into third
6. The method of any of the preceding claims, wherein the interaction site is associated with one or more interaction pages, wherein each interaction page represents a respective step of the multi-step communication flow, and wherein the one or more voice interaction are configured by a user of a content provider device through interactions with a development interface to include values for (i) one or more 20 parameters associated with the voice-interaction with the interaction site, (ii) one or more parameters associated with the visual-interaction with the interaction site, and (iii) one or more parameters associated with both the voice-interaction and the visualinteraction with the interaction site.
25
7. The method of claim 6, wherein the second instructions do not include instructions generated using the values for the one or more parameters associated with the visual-interaction with the interaction site, and wherein the third instructions do not
-502016246072 28 Mar 2019 include instructions generated using the values for the one or more parameters associated with the voice-interaction with the interaction site.
8. The method of any of the preceding claims, wherein the first instructions are
9. The method of any of the preceding claims, wherein the first instructions and the
10 in response to determining that the request from the communications device is for the voice-interaction with the interaction site:
retrieving the common interaction flow document;
translating the first instructions into second instructions, wherein the second instructions, when executed by the communications device, provide the one or 15 more steps through a voice interaction with a user of the communications device; and transmitting, to the communications device, the second instructions; and in response to determining that the request from the communications device is for the visual-interaction with the interaction site:
retrieving the common interaction flow document;
10 include instructions generated using the values for the one or more parameters associated with the voice-interaction with the interaction site.
10 translating the first instructions into second instructions , wherein the second instructions, when executed by the communications device, provide the one or more steps through a voice interaction with a user of the communications device; and transmitting, to the communications device, the second instructions; and
10. The method of any of the preceding claims, wherein the automated response system is an interactive voice response system.
10 second instructions are based on VoiceXML, and wherein the third instructions are based on HTML5.
10 5. The method of any of the preceding claims, further comprising storing data representing a communication between the communications device and the automated response system, the data including an indication of whether the communication is visual.
15
11. A system, comprising:
one or more processors and one or more non-transitory computer-readable storage devices storing instructions that when executed by the one or more processors cause the one or more processors to perform operations comprising:
20 receiving, from a communications device, a request to execute an interaction site for enabling communications to be exchanged between an automated response system and the communications device, wherein the interaction site specifies a multi-step communication flow between the communications device and the automated response system;
25 transmitting, to the automated response system, a request to execute the interaction site;
receiving, from the automated response system, a common interaction flow document including first instructions that, when executed, provide one or more
-512016246072 28 Mar 2019 steps of the multi-step communication flow between the communications device and the automated response system, wherein the common interaction flow document is for handling interactions received over different channels including voice interactions and visual interactions;
12. The system of claim 11, wherein transmitting the request to execute the interaction site comprises transmitting to the automated response system an indication
25 that indicates whether the request from the communications device is for the voiceinteraction with the interaction site or the visual-interaction with interaction site.
-522016246072 28 Mar 2019
13. The system of claim 11 or 12, wherein determining whether the request from the communications device is for the voice-interaction with the interaction site or the visualinteraction with the interaction site comprises:
determining whether the request is received from a voice channel or a data 5 channel;
in response to determining that the request is received from the voice channel, determining that the request is for the voice-interaction with the interaction site; and in response to determining that the request is received from the data channel, determining that the request is for the visual-interaction with the interaction site.
14. The system of claim 11. 12 or 13, wherein determining whether the request from the communications device is for the voice-interaction with the interaction site or the visual-interaction with the interaction site comprises:
determining that the request is based on hypertext markup language (HTML);
15 based on voice extensible markup language (VoiceXML), and wherein the third instructions are based on HTML5.
15 and in response to determining that the request is based on HTML, determining that the request from the communications device is for the visual-interaction with the interaction site.
20 15. The system of any of claims 11 to 14, wherein the operations further comprise storing data representing a communication between the communications device and the automated response system, the data including an indication of whether the communication is visual.
25
15 in response to determining that the request from the communications device is for the visual-interaction with the interaction site:
retrieving the common interaction flow document;
translating the first instructions into third instructions, wherein the third instructions, when executed by the communications device, provide the one or more 20 steps through a visual interaction with the user of the communications device; and transmitting, to the communications device, the third instructions.
15 site or the visual-interaction with interaction site.
16. The system of any of claims 11 to 15, wherein the interaction site is associated with one or more interaction pages, wherein each interaction page represents a respective step of the multi-step communication flow, and wherein the one or more voice interaction are configured by a user of a content provider device through
-532016246072 28 Mar 2019 interactions with a development interface to include values for (i) one or more parameters associated with the voice-interaction with the interaction site, (ii) one or more parameters associated with the visual-interaction with the interaction site, and (iii) one or more parameters associated with both the voice-interaction and the visual5 interaction with the interaction site.
17. The system of claim 16, wherein the second instructions do not include instructions generated using the values for the one or more parameters associated with the visual-interaction with the interaction site, and wherein the third instructions do not
18. The system of any of claims 11 to 17, wherein the first instructions are based on state chart extensible markup language (SCXML), wherein the second instructions are
19. The system of any of claims 11 to 18, wherein the first instructions and the second instructions are based on VoiceXML, and wherein the third instructions are
20 based on HTML5.
20. A non-transitory computer-readable storage device storing instructions that when executed by one or more processors, cause the one or more processors to perform operations comprising:
25 receiving, from a communications device, a request to execute an interaction site for enabling communications to be exchanged between an automated response system and the communications device, wherein the interaction site specifies a multi-542016246072 28 Mar 2019 step communication flow between the communications device and the automated response system;
transmitting, to the automated response system, a request to execute the interaction site;
20 translating the first instructions into third instructions , wherein the third instructions, when executed by the communications device, provide the one or more steps through a visual interaction with the user of the communications device; and, transmitting, to the communications device, the third instructions.
AU2016246072A 2015-04-10 2016-04-08 Visual interactive voice response system Active AU2016246072B2 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US14/683,894 2015-04-10
US14/683,894 US9467563B1 (en) 2015-04-10 2015-04-10 Visual interactive voice response system
PCT/US2016/026814 WO2016164842A1 (en) 2015-04-10 2016-04-08 Visual interactive voice response system

Publications (2)

Publication Number Publication Date
AU2016246072A1 AU2016246072A1 (en) 2017-11-30
AU2016246072B2 true AU2016246072B2 (en) 2019-04-18

Family

ID=57046355

Family Applications (1)

Application Number Title Priority Date Filing Date
AU2016246072A Active AU2016246072B2 (en) 2015-04-10 2016-04-08 Visual interactive voice response system

Country Status (7)

Country Link
US (2) US9467563B1 (en)
EP (1) EP3281395B1 (en)
KR (1) KR102107285B1 (en)
CN (1) CN107771389B (en)
AU (1) AU2016246072B2 (en)
CA (1) CA2992249C (en)
WO (1) WO2016164842A1 (en)

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9467563B1 (en) 2015-04-10 2016-10-11 Angel.Com Incorporated Visual interactive voice response system
CN105791538B (en) * 2016-02-17 2020-06-26 钉钉控股(开曼)有限公司 Prompting method and device
US10102856B2 (en) * 2017-01-20 2018-10-16 Essential Products, Inc. Assistant device with active and passive experience modes
US20180218729A1 (en) * 2017-01-31 2018-08-02 Interactive Intelligence Group, Inc. System and method for speech-based interaction resolution
US10970354B2 (en) * 2017-07-17 2021-04-06 Songtradr, Inc. Method for processing code and increasing website and client interaction speed
CN112073776B (en) * 2019-06-10 2022-06-24 海信视像科技股份有限公司 Voice control method and display device
CN110995942B (en) * 2019-12-06 2021-08-06 科大国创软件股份有限公司 Soft switch automatic calling method and system based on interface visualization
CN115567668A (en) * 2022-09-26 2023-01-03 中国建设银行股份有限公司 IVVR-based audio and video call processing method and device and computer equipment

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080162628A1 (en) * 2007-01-03 2008-07-03 Peter Hill Simultaneous visual and telephonic access to interactive information delivery

Family Cites Families (51)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6985478B2 (en) 1998-02-17 2006-01-10 Genesys Telecommunications Laboratories, Inc. Using XML expressed primitives for platform and system-independent call modeling
US6606611B1 (en) * 1999-02-27 2003-08-12 Emdadur Khan System and method for audio-only internet browsing using a standard telephone
US7082422B1 (en) 1999-03-23 2006-07-25 Microstrategy, Incorporated System and method for automatic transmission of audible on-line analytical processing system report output
US7457397B1 (en) 1999-08-24 2008-11-25 Microstrategy, Inc. Voice page directory system in a voice page creation and delivery system
US6792086B1 (en) * 1999-08-24 2004-09-14 Microstrategy, Inc. Voice network access provider system and method
US7340040B1 (en) 1999-09-13 2008-03-04 Microstrategy, Incorporated System and method for real-time, personalized, dynamic, interactive voice services for corporate-analysis related information
US6836537B1 (en) 1999-09-13 2004-12-28 Microstrategy Incorporated System and method for real-time, personalized, dynamic, interactive voice services for information related to existing travel schedule
US6964012B1 (en) 1999-09-13 2005-11-08 Microstrategy, Incorporated System and method for the creation and automatic deployment of personalized, dynamic and interactive voice services, including deployment through personalized broadcasts
US6940953B1 (en) 1999-09-13 2005-09-06 Microstrategy, Inc. System and method for the creation and automatic deployment of personalized, dynamic and interactive voice services including module for generating and formatting voice services
US6850603B1 (en) 1999-09-13 2005-02-01 Microstrategy, Incorporated System and method for the creation and automatic deployment of personalized dynamic and interactive voice services
US7197461B1 (en) 1999-09-13 2007-03-27 Microstrategy, Incorporated System and method for voice-enabled input for use in the creation and automatic deployment of personalized, dynamic, and interactive voice services
US7266181B1 (en) 1999-09-13 2007-09-04 Microstrategy, Incorporated System and method for the creation and automatic deployment of personalized dynamic and interactive voice services with integrated inbound and outbound voice services
US6768788B1 (en) 1999-09-13 2004-07-27 Microstrategy, Incorporated System and method for real-time, personalized, dynamic, interactive voice services for property-related information
US6885734B1 (en) 1999-09-13 2005-04-26 Microstrategy, Incorporated System and method for the creation and automatic deployment of personalized, dynamic and interactive inbound and outbound voice services, with real-time interactive voice database queries
US6829334B1 (en) 1999-09-13 2004-12-07 Microstrategy, Incorporated System and method for the creation and automatic deployment of personalized, dynamic and interactive voice services, with telephone-based service utilization and control
EP2136543A1 (en) * 2000-01-31 2009-12-23 Grape Technology Group, Inc. Communication assistance system and method
WO2001067225A2 (en) * 2000-03-06 2001-09-13 Kanisa Inc. A system and method for providing an intelligent multi-step dialog with a user
US6920425B1 (en) * 2000-05-16 2005-07-19 Nortel Networks Limited Visual interactive response system and method translated from interactive voice response for telephone utility
US6996800B2 (en) 2000-12-04 2006-02-07 International Business Machines Corporation MVC (model-view-controller) based multi-modal authoring tool and development environment
US7296226B2 (en) 2001-02-15 2007-11-13 Accenture Gmbh XML-based multi-format business services design pattern
CA2440807A1 (en) 2001-03-30 2002-10-10 British Telecommunications Public Limited Company Multi-modal interface
US7054939B2 (en) * 2001-06-28 2006-05-30 Bellsouth Intellectual Property Corportion Simultaneous visual and telephonic access to interactive information delivery
GB0129787D0 (en) 2001-12-13 2002-01-30 Hewlett Packard Co Method and system for collecting user-interest information regarding a picture
US7324633B2 (en) * 2002-05-30 2008-01-29 At&T Delaware Intellectual Property, Inc. Web integrated interactive voice response
US20040110487A1 (en) * 2002-12-09 2004-06-10 International Business Machines Corporation Wireless network access system
US7831693B2 (en) 2003-08-18 2010-11-09 Oracle America, Inc. Structured methodology and design patterns for web services
US7792254B2 (en) 2004-10-19 2010-09-07 Genesys Telecommunications Laboratories, Inc. System for distributing VXML capabilities for execution on client devices
CN1960408B (en) * 2005-10-31 2010-05-26 中国电信股份有限公司 Multimedia responsion method for interactive multimedia response system
US7996251B2 (en) * 2006-02-22 2011-08-09 24/7 Customer, Inc. System and method for customer requests and contact management
US20070233495A1 (en) * 2006-03-29 2007-10-04 International Business Machines Corporation Partially automated technology for converting a graphical interface to a speech-enabled interface
US8804694B2 (en) 2006-06-08 2014-08-12 At&T Intellectual Property Ii, L.P. Method and apparatus for invoking multimodal interaction in a VOIP call
US8086463B2 (en) 2006-09-12 2011-12-27 Nuance Communications, Inc. Dynamically generating a vocal help prompt in a multimodal application
US8200527B1 (en) 2007-04-25 2012-06-12 Convergys Cmg Utah, Inc. Method for prioritizing and presenting recommendations regarding organizaion's customer care capabilities
US8032379B2 (en) * 2007-06-20 2011-10-04 International Business Machines Corporation Creating and editing web 2.0 entries including voice enabled ones using a voice only interface
KR100936757B1 (en) * 2007-07-13 2010-01-15 (주) 콜게이트 Visual ARS Service System and Method Enabled by Mobile Terminal's Call Control Function
US8041575B2 (en) 2007-11-28 2011-10-18 International Business Machines Corporation System and method for enabling voice driven interactions among multiple IVR's, constituting a voice workflow
US9031982B2 (en) 2008-02-11 2015-05-12 Accenture Global Services Limited Multi-channel content modeling system
US8117538B2 (en) * 2008-12-19 2012-02-14 Genesys Telecommunications Laboratories, Inc. Method for dynamically converting voice XML scripts into other compatible markup language scripts based on required modality
CN101626436B (en) * 2009-06-22 2011-12-28 中兴通讯股份有限公司 Interactive video sound responding system and realizing method
US8249627B2 (en) * 2009-12-21 2012-08-21 Julia Olincy “I am driving/busy” automatic response system for mobile phones
US8406388B2 (en) * 2011-07-18 2013-03-26 Zvi Or-Bach Systems and methods for visual presentation and selection of IVR menu
US8898326B2 (en) 2010-02-05 2014-11-25 Oracle International Corporation Service deliver platform based support of interactions between next generation networks and legacy networks
US20110211679A1 (en) 2010-02-26 2011-09-01 Vladimir Mezhibovsky Voice Response Processing
US8655965B2 (en) * 2010-03-05 2014-02-18 Qualcomm Incorporated Automated messaging response in wireless communication systems
US8582727B2 (en) 2010-04-21 2013-11-12 Angel.Com Communication of information during a call
US8452597B2 (en) * 2011-09-30 2013-05-28 Google Inc. Systems and methods for continual speech recognition and detection in mobile computing devices
US9160844B2 (en) * 2012-08-06 2015-10-13 Angel.Com Incorporated Conversation assistant
US9148512B1 (en) * 2013-10-11 2015-09-29 Angel.Com Incorporated Routing user communications to agents
US20150256677A1 (en) 2014-03-07 2015-09-10 Genesys Telecommunications Laboratories, Inc. Conversation assistant
WO2015181775A1 (en) * 2014-05-29 2015-12-03 CafeX Communications, Ltd. Pushing web and application pages during video/audio calls
US9467563B1 (en) 2015-04-10 2016-10-11 Angel.Com Incorporated Visual interactive voice response system

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080162628A1 (en) * 2007-01-03 2008-07-03 Peter Hill Simultaneous visual and telephonic access to interactive information delivery

Also Published As

Publication number Publication date
EP3281395B1 (en) 2020-04-08
US20160301802A1 (en) 2016-10-13
KR102107285B1 (en) 2020-05-06
EP3281395A1 (en) 2018-02-14
EP3281395A4 (en) 2018-05-02
CA2992249C (en) 2020-07-14
US10270908B2 (en) 2019-04-23
KR20170141711A (en) 2017-12-26
CN107771389B (en) 2020-12-25
CA2992249A1 (en) 2016-10-13
CN107771389A (en) 2018-03-06
AU2016246072A1 (en) 2017-11-30
US20170118337A1 (en) 2017-04-27
US9467563B1 (en) 2016-10-11
WO2016164842A1 (en) 2016-10-13

Similar Documents

Publication Publication Date Title
AU2016246072B2 (en) Visual interactive voice response system
US10015315B2 (en) Call center builder platform
US10009463B2 (en) Multi-channel delivery platform
US10063701B2 (en) Custom grammars builder platform
US10359923B2 (en) Application builder platform
US10101974B2 (en) Contact center application creating using reusable program modules
US9479640B1 (en) Multimodal interactive voice response system
CA3084923C (en) Multi-channel delivery platform

Legal Events

Date Code Title Description
FGA Letters patent sealed or granted (standard patent)
HB Alteration of name in register

Owner name: GENESYS CLOUD SERVICES HOLDINGS II, LLC

Free format text: FORMER NAME(S): GREENEDEN U.S. HOLDINGS II LLC