US20070219803A1 - Method for creating and deploying system changes in a voice application system - Google Patents

Method for creating and deploying system changes in a voice application system Download PDF

Info

Publication number
US20070219803A1
US20070219803A1 US11/728,954 US72895407A US2007219803A1 US 20070219803 A1 US20070219803 A1 US 20070219803A1 US 72895407 A US72895407 A US 72895407A US 2007219803 A1 US2007219803 A1 US 2007219803A1
Authority
US
United States
Prior art keywords
voice
application
dialogue
edited
original
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/728,954
Inventor
Leo Chiu
Steven Ehrlich
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
HTC Corp
Original Assignee
Apptera Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US10/835,444 external-priority patent/US7817784B2/en
Application filed by Apptera Inc filed Critical Apptera Inc
Priority to US11/728,954 priority Critical patent/US20070219803A1/en
Publication of US20070219803A1 publication Critical patent/US20070219803A1/en
Assigned to HTC CORPORATION reassignment HTC CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: APPTERA, INC.
Assigned to APPTERA, INC. reassignment APPTERA, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: CHIU, LEO, EHRLICH, STEVEN CLIVE
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M3/00Automatic or semi-automatic exchanges
    • H04M3/42Systems providing special services or facilities to subscribers
    • H04M3/487Arrangements for providing information services, e.g. recorded voice services or time announcements
    • H04M3/493Interactive information services, e.g. directory enquiries ; Arrangements therefor, e.g. interactive voice response [IVR] systems or voice portals
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/08Configuration management of networks or network elements
    • H04L41/0803Configuration setting
    • H04L41/0813Configuration setting characterised by the conditions triggering a change of settings
    • H04L41/082Configuration setting characterised by the conditions triggering a change of settings the condition being updates or upgrades of network functionality
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/1066Session management
    • H04L65/1101Session protocols
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/34Network arrangements or protocols for supporting network services or applications involving the movement of software or configuration parameters 
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M3/00Automatic or semi-automatic exchanges
    • H04M3/42Systems providing special services or facilities to subscribers
    • H04M3/42136Administration or customisation of services
    • H04M3/42144Administration or customisation of services by service provider
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2203/00Aspects of automatic or semi-automatic exchanges
    • H04M2203/35Aspects of automatic or semi-automatic exchanges related to information services provided via a voice call
    • H04M2203/355Interactive dialogue design tools, features or methods
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M3/00Automatic or semi-automatic exchanges
    • H04M3/42Systems providing special services or facilities to subscribers
    • H04M3/487Arrangements for providing information services, e.g. recorded voice services or time announcements
    • H04M3/493Interactive information services, e.g. directory enquiries ; Arrangements therefor, e.g. interactive voice response [IVR] systems or voice portals
    • H04M3/4938Interactive information services, e.g. directory enquiries ; Arrangements therefor, e.g. interactive voice response [IVR] systems or voice portals comprising a voice browser which renders and interprets, e.g. VoiceXML

Definitions

  • the present invention is in the area of voice application software systems and pertains particularly to systems for developing and managing voice files linked for service to a voice application deployment system.
  • a speech application is one of the most challenging applications to develop, deploy and maintain in a communications environment.
  • Expertise required for developing and deploying a viable VXML application includes expertise in computer telephony integration (CTI) hardware and software or a data network telephony (DNT) equivalent, voice recognition software, text-to-speech software, and speech application logic.
  • CTI computer telephony integration
  • DNT data network telephony
  • VXML voice extensive markup language
  • the expertise require to develop a speech solution has been reduced somewhat.
  • VXML is a language that enables a software developer to focus on the application logic of the voice application without being required to configure underlying telephony components.
  • the developed voice application is run on a VXML interpreter that resides on and executes on the associated telephony system to deliver the solution.
  • Data network telephony (DNT) equivalents of voice delivery systems also exist like VoIP portals and the like.
  • voice prompts are sometimes prerecorded in a studio setting for a number of differing business scenarios and uploaded to the enterprise system server architecture for access and deployment during actual interaction with clients.
  • Pre-recording voice prompts instead of dynamically creating them through software and voice synthesis methods is many times performed when better sound quality, different languages, different voice types, or a combination of the above are desired for the presentation logic of a particular system.
  • voice prompts In very large enterprise architectures there may be many thousands of prerecorded voice prompts stored for use by a given voice application. Some of these may not be stored in a same centralized location.
  • voice file management will attest that managing such a large volume of voice prompts can be very complicated. For example, in prior-art systems management of voice prompts includes recording the prompts, managing identification of those prompts and manually referencing the required prompts in the application code used in developing the application logic for deployment of those prompts to a client interfacing system. There is much room for error in code referencing and actual development, recording, and sorting batches of voice files can be error prone and time consuming as well.
  • the inventor knows of a software interface for managing audio resources used in one or more voice applications.
  • the software interface includes a first portion for mapping the audio resources from storage to use-case positions in the one or more voice applications, a portion for accessing the audio resources according to the mapping information and for performing modifications a portion for creating new audio resources; and a portion for replication of modifications across distributed facilities.
  • a developer can modify or replace existing audio resources and replicate links to the application code of the applications that use them.
  • VXML-compliant and other types of voice systems may frequently need to be modified or updated, sometimes multiple times per day due to fast-paced business environments, rapidly evolving business models, special temporary product promotions, sales discounts and so on. For example, if a product line goes obsolete, existing voice prompts related to that product line that are operational in a deployed voice application may need to be modified, replaced or simply deleted. Moreover, configuration settings of a voice application interaction system may also need to be updated or modified from time to time due to addition of new hardware, software, and so on.
  • the software application mentioned above as known to the inventor for managing audio resources enables frequent modifications of existing voice applications in a much improved and efficient manner than in the current art.
  • the running voice application is typically suspended from service while the changes are implemented.
  • Shutting down service for even a temporary period can result in monetary loss that can be significant depending on the amount of time the system will be shut down.
  • a backup system may be deployed while the primary system is being reconfigured.
  • this approach requires more resources that would be required to run one application.
  • FIG. 1 is a logical overview of a voice interaction server and voice prompt data store according to prior-art.
  • FIG. 2 is a block diagram illustrating voice prompt development and linking to a voice prompt application according to prior art.
  • FIG. 3 is a block diagram illustrating a voice prompt development and management system according to an embodiment of the present invention.
  • FIG. 4 illustrates an interactive screen for a voice application resource management application according to an embodiment of the present invention.
  • FIG. 5 illustrates an interactive screen having audio resource details and dependencies according to an embodiment of the present invention.
  • FIG. 6 illustrates an interactive screen for an audio resource manager illustrating further details and options for editing and management according to an embodiment of the present invention.
  • FIG. 7 is a process flow diagram illustrating steps for editing or replacing an existing audio resource and replicating the resource to distributed storage facilities.
  • FIG. 8 is architectural overview of a communications network wherein automated voice application system configuration is practiced according to an embodiment of the present invention.
  • FIG. 9 is an exemplary screenshot illustrating application of modifications to a voice dialog according to an embodiment of the present invention.
  • FIG. 10 is a block diagram illustrating components of an automated voice application configuration application according to an embodiment of the present invention.
  • FIG. 11 is a process flow chart illustrating steps for receiving and implementing a change-order according to an embodiment of the present invention.
  • the inventor provides a system for managing voice prompts in a voice application system. Detail about methods, apparatus and the system as a whole are described in enabling detail below.
  • FIG. 1 is a logical overview of a voice interaction server and voice prompt data store according to prior art.
  • FIG. 2 is a block diagram illustrating voice prompt development and linking to a voice prompt application according to prior art.
  • a voice application system 100 includes a developer 101 , a voice file storage medium 102 , a voice portal (telephony, IVR) 103 , and one of possibly hundreds or thousands of receiving devices 106 .
  • IVR voice portal
  • Device 106 may be a LAN-line telephone, a cellular wireless, or any other communication device that supports voice and text communication over a network.
  • device 106 is a plane old telephone service (POTS) telephone.
  • POTS plane old telephone service
  • Device 106 has access through a typical telephone service network, represented herein by a voice link 110 , to a voice system 103 , which in this example is a standard telephony IVR system.
  • IVR system 103 is the customer access point for callers (device 106 ) to any enterprise hosting or leasing the system.
  • IVR 103 has a database/resource adapter 109 for enabling access to off-system data.
  • IVR also has voice applications 108 accessible therein and adapted to provide customer interaction and call flow management. Applications 108 include the capabilities of prompting a customer, taking input from a customer and playing prompts back to the customer depending on the input received.
  • Telephony hardware and software 107 includes the hardware and software that may be necessary for customer connection and management of call control protocols.
  • IVR 103 may be a telephony switch enhanced as a customer interface by applications 108 .
  • Voice prompts executed within system 103 may include only prerecorded prompts.
  • a DNT equivalent may use both prerecorded prompts and XML-based scripts that are interpreted by a text-to-speech engine and played using a sampled voice.
  • IVR system 103 has access to a voice file data store 102 via a data link 104 , which may be a high-speed fiber optics link or another suitable data carrier many of which are known and available.
  • Data store 102 is adapted to contain prerecorded voice files, sometimes referred to as prompts. Prompts are maintained, in this example, in a section 113 of data store 102 adapted for the purpose of storing them.
  • a voice file index 112 is illustrated and provides a means for searching store section 113 to access files for transmission over link 104 to IVR system 103 to be played by one of applications 108 during interaction with a client.
  • IVR system 102 is a distributed system such as to a telephony switch location in a public switched telephone network (PSTN) and therefore is not equipped to store many voice files, which take up considerable storage space if they are high quality recordings.
  • PSTN public switched telephone network
  • Data store 111 has a developer/enterprise interface 111 for enabling developers such as developer 101 access for revising existing voice files and storing new and deleting old voice files from the data store.
  • Developer 101 may create voice applications and link stored voice files to the application code for each voice application created and deployed.
  • the voice files themselves are created in a separate studio from script provided by the developer.
  • the studio has to manage the files and present them to the developer in a fashion that the developer can manipulate in an organized fashion. As the number of individual prerecorded files increases, so does the complexity of managing those prerecorded files.
  • voice files are recorded from script. Therefore, for a particular application developer 101 creates enterprise scripts 202 and sends them out to a studio ( 200 ) to be recorded. An operator within studio 200 receives scripts 202 and creates recorded voice files 203 .
  • the files are single segments, some of which may be strategically linked together in a voice application to play as a single voice prompt to a client as part of a dialog executed from the point of IVR 103 , for example.
  • the enterprise must insure that voice files 203 are all current and correct and that the parent application has all of the appropriate linking in the appropriate junctions so that the files may be called up correctly during execution.
  • Developer 101 uploads files 203 when complete to data store 102 and the related application may also be uploaded to data store 102 .
  • a specific application needs to be run at a customer interface, it may be distributed without the voice files to the point of interface, in this case IVR 103 .
  • IVR 103 There may be many separate applications or sub-dialogs that use the same individual voice files. Often there will be many instances of the same voice file stored in data store 102 but linked to separate applications that use the same prompt in some sequence.
  • FIG. 3 is an expanded view of IVR 103 of FIG. 2 illustrating a main dialog and sub-dialogs of a voice application according to prior art.
  • a main dialog 300 includes a static interactive menu 301 that is executed as part of the application logic for every client that calls in.
  • a client may provide input 302 , typically in the form of voice for systems equipped with voice recognition technology.
  • a system response 303 is played according to input 302 .
  • System response 303 may include as options, sub-dialogs 304 ( a - n ).
  • Sub-dialogs 304 ( a - n ) may link any number of prompts, or voice files 305 ( a - n ) illustrated logically herein for each illustrated sub-dialog.
  • prompt 305 b is used in sub-dialog 304 a and in sub-dialog 304 b.
  • Prompt 305 c is used in all three sub-dialogs illustrated.
  • Prompt 305 a is used in sub-dialog 304 b and in sub-dialog 304 b.
  • Prompts are created at the time of application creation and deployment. Therefore prompts 305 b, c, and j are stored in separate versions and locations for each voice application.
  • FIG. 4 illustrates an interactive screen 400 for a voice application resource management application according to an embodiment of the present invention.
  • Screen 400 is a GUI portion of a software application that enables a developer to create and manage resources used in voice applications.
  • Resources include both audio resources and application scripts that may be voice synthesized.
  • the inventor focuses on management of audio resources, which in this case, include voice file or prompt management in the context of one or more voice file applications.
  • Screen 400 takes the form of a Web browser type interface and can be used to access remote resources over a local area network (LAN), wide area network (WAN), or a metropolitan area network (MAN).
  • LAN local area network
  • WAN wide area network
  • MAN metropolitan area network
  • a developer operating through screen 400 is accessing a local Intranet.
  • Screen 400 has a toolbar link 403 that is labeled workspace.
  • Link 403 is adapted to open, upon invocation, a second window or changes the primary window to provide an area for working and audio management and creation tools for creating and working with audio files and transcripts or scripts.
  • Screen 400 has a toolbar link 404 that is labeled application.
  • Link 404 is adapted to open, upon invocation, a second window or changes the primary window to provide an area for displaying and working with voice application code and provides audio resource linking capability.
  • Screen 400 also has a toolbar link for enabling an administration view of all activity.
  • Screen 400 has additional toolbar links 406 adapted for navigating to different windows generally defined by label. Reading from left to right in toolbar options 406 , there is Audio, Grammar, Data Adapter, and Thesaurus.
  • the option Audio enables a user to view all audio-related resources.
  • the option Grammar enables a user to view all grammar-related resources.
  • the option Data Adapter enables a user to view all of the available adapters used with data sources, including adapters that might exist between disparate data formats.
  • the option Thesaurus is self-descriptive.
  • a developer has accessed the audio resource view, which provides in window 409 an interactive data list 411 of existing audio resources currently available in the system.
  • List 411 is divided into two columns a column 408 labeled “name” and a column 410 labeled “transcript”.
  • An audio speaker icon next to each list item indicates the item is an audio resource and enable a developer to.
  • Each audio resource is associated with the appropriate transcript of the resource as illustrated in column 410 . Reading from top to bottom in column 410 for the audio resource “howmuch” the transcript is “How much do you wish to transfer?”.
  • a scroll function may be provided to scroll a long transcript associated with an audio resource.
  • the transcript is “Your balance is [ ].
  • the brackets enclose a variable used in a voice system prompt response to a client input interpreted by the system.
  • separate views of directory 411 may be provided in different languages.
  • separate views of directory 411 may be provided for the same resources recorded using different voice talents.
  • voice files that are contextually the same, but are recorded using different voice talents and or languages, those files may be stored together and versioned according to language and talent.
  • Window 409 can be scrollable to reach any audio resources not viewable in the immediate screen area.
  • a left-side navigation window may be provided that contains both audio resource and grammar resource indexes 401 and 402 respectively to enable quick navigation through the lists.
  • a resource search function 411 is also provided in this example to enable keyword searching of audio and grammar resources.
  • Screen 400 has operational connectivity to a data store or stores used to where house the audio and grammar resources and, in some cases, the complete voice applications. Management actions initiated through the interface are applied automatically to the resources and voice applications.
  • a set of icons 407 defines additional interactive options for initiating immediate actions or views. For example, accounting from left to right a first icon enables creation of a new audio resource from a written script. Invocation of this icon brings up audio recording and editing tools that can be used to create new audio voice files and that can be used to edit or version existing audio voice files.
  • a second icon is a recycle bin for deleting audio resources.
  • a third icon in grouping 407 enables an audio resource to be copied.
  • a fourth icon in grouping 407 enables a developer to view a dependency tree illustrating if where and when the audio file is used in one or more voice dialogs. The remaining two icons are upload and download icons enabling the movement of audio resources from local to remote and from remote to local storage devices.
  • the functions of creating voice files and linking them to voice applications can be coordinated through interface 400 by enabling an author of voice files password protected local or remote access for downloading enterprise scripts and for uploading new voice files to the enterprise voice file database.
  • an operator calls up a next screen illustrating more detail about the resources and further options for editing and management as will be described below.
  • Screen 400 in this example, has and audio index display area 401 and a grammar display index area 402 strategically located in a left scrollable sub-window of screen 400 .
  • a grammar display index area 402 strategically located in a left scrollable sub-window of screen 400 .
  • the same resource may be highlighted in the associated index 401 or 402 depending on the type of resource listed.
  • FIG. 5 is illustrates an interactive screen 500 showing audio resource details and dependencies according to an embodiment of the present invention.
  • Screen 500 has a scrollable main window 501 that is adapted to display further details about audio resources previously selected for view. Previous options 406 remain displayed in screen 500 .
  • each resource selected in screen 400 is displayed in list form.
  • audio resource 504 has a resource name “howmuch”.
  • the resource 504 is categorized according to Dialog, Dialog type, and where the resource is used in existing voice applications.
  • the dialog reference is “How Much”
  • the resource type is a dialog, and the resource is used in a specified dialog prompt. Only one dependency is listed for audio resource 504 , however all dependencies (if more than one) will be listed.
  • Resource 505 “mainmenu” has dependency to two main menus associated with dialogs. In the first listing the resource is used in a standard prompt used in the first listed dialog of the first listed main menu. In the second row it is illustrated that the same audio resource also is used in a nomatch prompt used in a specified dialog associated with the second listed main menu.
  • a nomatch prompt is one where the system does not have to match any data provided in a response to the prompt.
  • a noinput prompt is one where no input is solicited by the prompt. It is noted herein that for a general application prompt definitions may vary widely according to voice application protocols and constructs used.
  • the dependencies listed for resource 505 may be associated with entirely different voice applications used by the same enterprise. They may also reflect dependency of the resource to two separate menus and dialogs of a same voice application.
  • No specific ID information is illustrated in this example, but may be assumed to be present. For example, there may be rows and columns added for displaying a URL or URI path to the instance of the resource identified. Project Name, Project ID, Project Date, Recording Status (new vs. recorded), Voice Talent, and Audio Format are just some of the detailed information that may be made available in window 501 . There may be a row or column added for provision of a general description of the resource including size, file format type, general content, and so on.
  • Resource 506 “yourbalance” is listed with no dependencies found for the resource. This may be because it is a newly uploaded resource that has not yet been linked to voice application code. It may be that it is a discarded resource that is still physically maintained in a database for possible future use. The lack of information tells the operator that the resource is currently not being used anywhere in the system.
  • Screen 500 in this example, has audio index display area 401 and a grammar display index area 402 strategically located in a left scrollable sub-window of screen 500 as described with reference to screen 400 of FIG. 4 above.
  • a grammar display index area 402 strategically located in a left scrollable sub-window of screen 500 as described with reference to screen 400 of FIG. 4 above.
  • the same resource may be highlighted in the associated index 401 or 402 depending on the type of resource listed.
  • FIG. 6 illustrates an interactive screen 600 of an audio resource manager illustrating further details and options for editing and management according to an embodiment of the present invention.
  • Screen 600 enables a developer to edit existing voice files and to create new voice files.
  • a dialog tree window 602 is provided and is adapted to list all of the existing prompts and voice files linked to dialogs in voice applications. The information is, in a preferred embodiment, navigable using a convenient directory and file system format. Any voice prompt or audio resource displayed in the main window 601 is highlighted in the tree of window 602 .
  • a developer can download a batch of audio resources (files) from a studio remotely, or from local storage and can link those into an existing dialog, or can create a new dialog using the new files.
  • the process leverages an existing database program such as MS ExcelTM for versioning and keeping track of voice prompts dialogs, sub-dialogs, and other options executed during voice interaction.
  • a developer can navigate using the mapping feature through all of the voice application dialogs referencing any selected voice files.
  • the dialogs can be presented in descending or ascending orders according to some criteria specified like date, number of use positions, or some other hierarchical specification.
  • a developer accessing an audio resource may also have access to any associated reference files like coaching notes, contextual notes, voice talent preferences, language preferences, and pronunciation nuances for different regions.
  • multiple links do not have to be created to replace an audio resource used in multiple dialog prompts of one or more voice applications. For example, after modifying a single voice file, one click may cause the link to the stored resource to be updated across all instances of the file in all existing applications.
  • replication may be ordered such that the modified file is automatically replicated to all of the appropriate storage sites for local access. In this case, the resource linking is updated to each voice application using the file according to the replication location for that application.
  • Screen 600 illustrates a prompt 604 being developed or modified.
  • the prompt in this example is named “Is that correct?” and has variable input fields of City and State.
  • the prompt 604 combines audio files to recite “You said [City: State]:: If that is correct, say Yes: If in correct, Say No:
  • the prompt may be used in more than one dialog in more than one voice application.
  • the prompt may incorporate more than one individual prerecorded voice file.
  • a window 605 contains segment information associated with the prompt “Is that correct?” such as the variable City and State and the optional transcripts (actual transcripts of voice files). New voice files and transcripts describing new cities and states may be added and automatically liked to all of the appropriate prompt segments used in all dialogs and applications.
  • audio voice files of a same content definition but prerecorded in one or more different languages and/or voice talents will be stored as separate versions of the file.
  • automated voice translation utilities can be used to translate an English voice file into a Spanish voice file, for example, on the fly as the file is being accessed and utilized in an application. Therefore, in a more advanced embodiment multiple physical prerecorded voice files do not have to be maintained.
  • Screen 600 has a set of options 603 for viewing creating or editing prompts, rules, nomatch prompts, and no-input prompts. Options for help, viewing processor details, help with grammar, and properties are also provided within option set 603 .
  • Workspace provides input screen or windows for adding new material and changes. The workspace windows can be in the form of an excel worksheet as previously described.
  • linking voice files to prompts in application can be managed across multiple servers in a distributed network environment.
  • Voice files, associated transcripts, prompt positions, dialog positions, and application associations are all automatically applied for the editor eliminating prior-art practice of re-linking the new resources in the application code.
  • Other options not illustrated in this example may also be provided without departing from the spirit and scope of the present invention. For example, when a voice file used in several places has been modified, the editor may not want the exact version to be automatically placed in all use instances. In this case, the previous file is retained and the editor simply calls up a list of the use positions and selects only the positions that the new file applies to. The system then applies the new linking for only the selected prompts and dialogs. The old file retains the linking to the appropriate instances where no modification was required.
  • voice file replication across distributed storage systems is automated for multiple distributed IVR systems or VXML portals. For example, if a developer makes changes to voice files in one storage facility and links those changes to all known instances of their use at other client access points, which may be widely distributed, then the distributed instances may automatically order replication of the appropriate audio resources from the first storage facility to all of the other required storage areas. Therefore, for voice applications that are maintained at local client-access facilities of a large enterprise that rely on local storage of prerecorded files can, after receiving notification of voice file linking to a new file or files can execute and order to retrieve those files from the original storage location and deposit them into their local stores for immediate access. The linking then is used as a road map to insure that all distributed sites using the same applications have access to all of the required files.
  • audio resource editing can be performed at any network address wherein the changes can be automatically applied to all distributed facilities over a WAN.
  • FIG. 7 is a process flow diagram 700 illustrating steps for editing or replacing an existing audio resource and replicating the resource to distributed storage facilities.
  • the developer selects an audio resource for edit or replacement. The selection can be based on a search action for a specific audio resource or from navigation through a voice application dialog menu tree.
  • dialogs that reference the selected audio resource are displayed.
  • the developer may select the dialogs that will use the edited or replacement resource by marking or highlighting those listed dialogs. In one embodiment all dialogs may be selected. The exact number of dialogs selected will depend on the enterprise purpose of the edit or replacement.
  • the developer edits and tests the new resource, or creates an entirely new replacement resource.
  • the developer saves the final tested version of the resource.
  • the version saved is automatically replicated to the appropriate storage locations referenced by the dialogs selected in step 703 .
  • steps 702 , and step 706 are automated results of the previous actions performed.
  • the methods and apparatus of the present invention can be applied on a local network using a central or distributed storage system as well as over a WAN using distributed or central storage. Management can be performed locally or remotely, such as by logging onto the Internet or an Intranet to access the software using password protection and/or other authentication procedures.
  • the methods and apparatus of the present invention greatly enhance and streamline voice application development and deployment and according to the embodiments described, can be applied over a variety of different network architectures including DNT and POTS implementations.
  • a software routine is provided that is capable of receiving a configuration package and of implementing the package at a point of voice interaction in order to effect system changes and voice application changes without suspending a system or application that is running and in the process of interaction with clients.
  • FIG. 8 is architectural overview of a communications network 800 wherein automated voice application system configuration is practiced according to an embodiment of the present invention.
  • Communications network 800 encompasses a wide-area-network (WAN) 801 , a public-switched-telephone-network (PSTN) 802 , and a communications host illustrated herein as an enterprise 803 .
  • WAN wide-area-network
  • PSTN public-switched-telephone-network
  • Enterprise 803 may be any type of enterprise that provides services to clients, which are accessible to a call-in center or department. Enterprise 803 , in this example, maintains voice interaction access points to voice services. Enterprise 803 may be assumed to contain a communications-center type environment wherein service agents interact with clients calling into the enterprise.
  • Enterprise 803 has a local-area-network (LAN) 820 provided therein and adapted for supporting a plurality of agent-operated workstations for communication and data sharing.
  • LAN 820 has communications access to WAN 801 and to PSTN 802 .
  • a central telephony switch (CS) 821 is provided within enterprise 803 and is adapted to receive calls routed thereto from PSTN 802 via a telephony trunk branch 817 from a local switch in the network illustrated herein as switch (LS) 804 .
  • LS 804 may be a private-branch type of exchange (PBX), and automated-call-distributor (ACD), or any other type of telephone switch capable of running calls.
  • PBX private-branch type of exchange
  • ACD automated-call-distributor
  • CS 821 has an interactive voice system peripheral (VS) 822 connected thereto by a CTI link.
  • VS 822 also has connection to LAN 820 .
  • VS 822 is adapted to interact with callers routed CS 821 according to voice application dialogs therein.
  • VS 822 may be an IVR system or a voice recognition system (VRS) without departing from the spirit and scope of the present invention.
  • VS 822 is a point of deployment for voice applications used for client interaction.
  • incoming calls routed to CS 821 from LS 800 from within PSTN 802 are illustrated as calls 805 incoming into LS 804 from anywhere within PSTN 805 .
  • VAS 824 has a voice application server (VAS) 824 provided therein and connected to LAN 820 .
  • VAS 824 is adapted for storing and serving voice applications created by an administrator (ADMN) 823 represented herein by a computer icon also shown connected to LAN 820 .
  • Administrator 823 uses a client software application (AS) 825 to create voice applications and manage voice files, voice prompts, and voice dialogs of those applications.
  • ADMN administrator
  • AS client software application
  • system 822 stores voice applications locally (storage not shown). In another embodiment of the present invention system 822 retrieves voice applications from system 824 over LAN 820 when those applications are required in interaction with clients.
  • AS 825 installed on workstation 823 is analogous to an application described further above respect to screenshots 400 , 500 , and 600 of FIGS. 4, 5 , and 6 respectively.
  • AS 825 is enhanced according to an embodiment of the present invention with a utility for enabling configuration and one touch deployment of voice application or system modification updates to voice applications or settings active at VS 822 .
  • updates created and deployed from workstation 823 are applied to voice applications while those applications are active in interaction without a requirement for shutting down or suspending those applications from service.
  • Voice application server 824 has connection to WAN 801 via a WAN access line 814 .
  • WAN 801 may be the well-known Internet, an Intranet, or a corporate WAN, among other possibilities.
  • LAN access line 814 may be a 24/7 connection or a connection through a network service provider.
  • WAN 801 has a network backbone 812 extending there through, which represents all of the lines, equipment, and access points making up the entire WAN as a whole.
  • Backbone 812 has a voice system (VS) 813 connected thereto, which represents a data-network-telephony (DNT) version of VS 822 .
  • System 813 uses voice applications to interact with clients accessing the system from anywhere in WAN 801 or any connected sub networks. It is noted herein, that networks 802 and 801 are bridged to gather for communication via a gateway 816 .
  • Gateway 816 is adapted translating telephony protocols into data network protocols and in reverse order enabling, for example, IP telephony callers to place calls to PSTN destinations, and PSTN telephony callers to place calls to WAN destinations.
  • gateway 816 may be an SS-7 Bell core system, or some other like system. Therefore, it is possible for PSTN callers to access voice interaction provided by system 813 and for WAN callers to access voice interaction provided by system 822 .
  • a remote administrator is illustrated in this example as a remote administrator 818 .
  • Administrator 818 may be operating from a remote office, from a home, or from any physical location providing telephone and network-access services.
  • a personal computer icon representing a workstation 819 further defines administrator 818 .
  • Workstation 819 is analogous in this embodiment to workstation 823 except that it is a remote workstation and not LAN-connected in this example.
  • Workstation 819 has a software application 825 a provided thereto, which is analogous to application 825 installed on workstation 823 within enterprise 803 .
  • Voice systems 822 and 813 have instances of a configuration order routine (COR), 826 for VS 822 , and 826 a for VS 813 installed thereon.
  • COR ( 826 , 826 a ) is adapted to except a configuration order package from AS 825 and/or AS 825 a respectively.
  • COR ( 826 , 826 a ) excepts and implements configuration orders created by administrators 823 or 819 and automatically applies those configuration orders to their respective voice systems.
  • administrator 823 utilizes application software 825 create necessary updates to existing voice applications including any required settings changes.
  • Voice application server 824 contains the actual voice applications in this case, which may be served to voice system 822 when required. In one embodiment however, voice system 822 may store voice applications for immediate access.
  • administrator 823 may initiate a one-touch deployment action that causes a change-order to be implemented by change-order routine 826 running in VS 822 . It is noted herein that a change-order for a voice application that is running may automatically extract and implement itself while the application is still running. A change-order may also be implemented to an application that is not currently running without departing from the spirit and scope of the present invention.
  • VS 822 When VS 822 receives a change-order from administrator 823 , application 826 executes and implements the change-order.
  • application 826 In the case of a running application, there may be a plurality of callers queued for different dialog prompts or prompt sequences of the same application. In this case, COR 826 monitors the state of the running application and implements the changes so that they do not negatively affect caller interaction with the application. More detail about how this is accomplished is provided later in this specification.
  • Remote administrator 819 may also create and implement change-orders to applications running in voice system 822 from a remote location. For example, utilizing AS 825 a, administrator 819 may connect to ISP 809 through LS 804 via trunk 806 and trunk branch 808 . ISP 809 may then connect administrator 819 to backbone 812 , from where VS 824 is accessible via network line 814 . Administrator 819 may therefore perform any of the types of edits or changes to applications running in VS 822 or to any settings of VS 822 that administrator 823 could configure for the same. Moreover, administrators 823 and 819 may generate updates for any voice applications running on voice system 813 connected to backbone 812 in WAN 801 .
  • Calls 805 may represent PSTN callers accessing CS 821 through trunk 806 and trunk branch 817 . Calls 805 may also include callers operating computers accessing VS 813 through ISP 809 via trunk branch 808 and network line 810 , or through gateway 816 via trunk branch 807 and network line 815 . Although the architecture in this example illustrates tethered access, callers 805 may also represent wireless users.
  • FIG. 9 is an exemplary interactive screen 900 illustrating application of modifications to a voice dialog according to an embodiment of the present invention.
  • Screen 900 illustrates capability for creating a change-order or update to voice application dialog in this example.
  • Screen 900 is a functional part of AS 825 or 825 a described above with reference to FIG. 8 .
  • Screenshot 900 in a preferred embodiment, stems from the same parent application hosting interactive screens 400 , 500 , and 600 , described further above.
  • Interactive screen 900 contains a workspace 902 , and a workspace 903 .
  • Space 902 contains a portion 904 of a dialog D-01 (logical representation only) illustrated in expanded view as a dialog 901 , which is accessible from a dialog menu illustrated at far left of screen 900 .
  • a dialog search box is provided for locating any particular dialog that needs to be updated.
  • dialog portion 904 is illustrated in the form of an original configuration.
  • a prompt 906 and a prompt 908 of dialog portion 904 will be affected by an update.
  • Dialog portion 900 is illustrated within workspace 903 as an edited version 905 .
  • Workspace 903 is a new configuration workspace.
  • Prompt 906 in workspace 902 is to be replaced.
  • the affected prompt is illustrated as a dotted rectangle containing an R signifying replacement.
  • prompt 906 is replaced with a prompt sequence 907 .
  • Sequence 907 contains three prompts labeled A signifying addition.
  • Prompt 908 from workspace 902 is illustrated as a deleted prompt 909 in workspace 903 (dotted rectangle D).
  • the new configuration 905 can be “saved-to-file” by activating a save button 910 , or can be saved and deployed by activating a deploy button 911 .
  • a reset button is also provided for resetting new configuration 905 to the form of the original the configuration 904 .
  • Interactive options for selecting prompts and for selecting attributes are provided for locating the appropriate new files link to the dialog.
  • Each workspace 902 and 903 has a prompt-view option enabling an administrator to select any prompt in the tree and expand that prompt for play-back purposes or for viewing transcripts, author data, and so on.
  • selecting the deploy option 911 causes the update package to be deployed to the appropriate VS system (if stored therein) or to the VAS if the application is executed from such a server.
  • the exact point of access for any voice system will depend on the purpose and design of the system. For example, referring back to FIG. 8 , if a voice system and switch are provided locally within an enterprise, then the actual voice applications may be served to clients through the voice system, the application hosted on a separate machine, but called in to service when needed.
  • VS 824 distributes the voice applications to the respective interaction points or hosts, especially if the interaction host machine is remote.
  • FIG. 10 is a block diagram illustrating components of automated voice application configuration routine ( 826 , 826 a ) according to an embodiment of the present invention.
  • Application 826 contains several components that enable automated configuration of updates or edits to voice applications that may be in the process of assisting clients.
  • Application 826 has a server port interface 1000 adapted to enable the application to detect when a change-order or update has arrived at the voice system.
  • a host machine running application 826 in a preferred embodiment, will have a cache memory or data queue adapted to contain incoming updates to voice applications, some of which may be running when the updates have arrived.
  • Application 826 has a scheduler component provided therein and adapted to receive change-orders from a cache memory and schedule those change-orders for task loading. It is noted herein that a change-order may have its own schedule for task loading. In this case scheduler 1002 parses the schedule of the change-order and will not load the order until the correct time has arrived.
  • Application 826 has a task loader 1003 provided therein and adapted to accept change-orders from scheduler 1002 for immediate implementation.
  • application 826 receives change-orders that include both instructions and the actual files required to complete the edits.
  • application 826 receives only the instructions, perhaps in the form of an object map or bitmap image, wherein the actual files are preloaded in identifiable fashion into a database containing the original files of the voice application or voice system settings.
  • the actual implementation will depend on whether the voice files used to update the application are stored locally (within the VS) or are accessed from a separate machine such as a VAS.
  • Application 826 has a voice application (VA) locator 1004 provided therein, and adapted to find, in the case of voice application update, the correct application that will be updated. It is possible that the application being updated is not in use currently. It is also possible that the application being updated is currently in use. In either instance, VA locator 1004 is responsible for finding the location of the application and its base files.
  • VA locator 1004 is responsible for finding the location of the application and its base files.
  • VA locator 1004 has connection to a database or server base interface 1006 provided therein and adapted to enable VA locator 1004 to communicate externally from the host system or VS. Therefore, if a particular voice application is being stored on a voice application server separate from voice system that uses the interaction, the voice application locator running on the voice system can locate correct application on the external machine.
  • Application 826 has a voice application (VA) state monitor 1005 provided therein and adapted to monitor state of any voice application identified by VA locator 1004 that is currently running and serving clients during the time of update.
  • State monitor 1005 has connection to a dialog controller interface 1009 .
  • a dialog controller is used by the voice system to execute a voice application.
  • the dialog controller manages the caller access and dialog flow of any voice application in use by the system and therefore has state information regarding the number of clients interacting with the application and their positions in the dialog hierarchy.
  • Application 826 has a sub-task scheduler/execution module 1007 provided therein, and adapted to execute a change-order task according to instructions provided by VA state monitor 1005 .
  • Module 1007 contains an orphan controller 1008 .
  • Orphan controller 1008 is adapted to maintain a functioning state in a voice application of certain prompts or prompt sequences that are to be deleted or replaced with new files used by a new configuration.
  • a new configuration is an object tree wherein the objects are prompts and prompt sequences.
  • the voice application that is to be modified has a similar object tree.
  • the objects or nodes are links to the actual files that are applied in voice interaction.
  • Module 1007 cooperates with VA state monitor 1005 to perform a change-order to a voice application using orphan controller 1008 to maintain functional orphans until all of the new objects are in place and callers are cleared from the orphan tree.
  • the voice application being modified continues to function as a backup application while it is being modified.
  • Replacement files and code modules associated with the change-order are, in a preferred embodiment, available in a same data store and memory partition that the original application files and code reside having been loaded therein either from cache or directly.
  • the files representing changes may be preloaded into the same storage hosting the old files such that as change-order is implemented by application 826 the change files are caused to take the place of the original files as required.
  • module 1007 works with VA state monitor 1005 , which in turn has connection to the application dialog controller, which in turn has connection to the telephony hardware facilitating client connection to voice applications. Therefore module 1007 can apply changes to the application and maintain orphan state until all of the accessing callers are interacting with the new configuration in a seamless matter. At that point the orphans (old files and settings) may be purged from the system.
  • Application 826 has a task state/completion notification module 1010 provided therein and adapted to send notification of the completed task to the task author or administrator through server port interface 1000 .
  • Module 1010 also has connection to change-order cache interface 1001 for the purpose of purging the cache of any data associated with a task that has been completed successfully.
  • module 1010 may send, through interface 1000 , an error notification or an advisory notification related to a change-order task that for some reason has not loaded successfully or that cannot be implemented efficiently. In the latter case, it may be that due to an unusually heavy call load using an existing application a change-order may be better scheduled during a time when there are not as many clients accessing the system.
  • this is not required in practice the present invention as during change-order implementation, nodes are treated individually in terms of caller access and as long as the new changes are implemented from the bottom up callers may be transferred from an orphan, for example, to a new object in a dialog tree until such time that that orphan may be replaced or deleted and so on.
  • Application 826 may be provided as a software application or routine that takes instruction directly from the change-orders it receives. In one embodiment of the present invention application 826 may be provided to run on a piece of dedicated hardware as firmware, the hardware having connection to the voice system. There are much possible variant architecture that may be used without departing from the spirit and scope of the present invention.
  • FIG. 11 is a process flow chart 1100 illustrating steps or receiving and implementing a change the according to an embodiment of the present invention.
  • a change-order is received by the system.
  • the actual files of the change-order may be cached in a cache memory and the change-order instructions, which in one embodiment are of the form of an executable bitmap or object model, are loaded into a task loader analogous to loader 1003 of FIG. 10 for processing.
  • the system locates the voice application that is the target of the change-order.
  • the target voice application may not be in current use. In this case, the changes may be implemented without concern for active state of the application interaction with clients. In another embodiment the target voice application may be currently in use with one or more of callers interacting with it.
  • the system prepares for execution of the change implementation task.
  • the current running state of the voice application is acquired. This information may include the total number of callers currently interacting with the application and their current positions of interaction with the application. Step 1104 is an ongoing step meaning that the system constantly receives current application state with respect to the number of callers and caller position in the dialog flow of the application.
  • step 1105 execution of the change-order begins.
  • step 1106 any orphans in the old application are identified and maintained from the top or root node of the application down the hierarchy until they are idle or not in a current state of access from one or more clients.
  • step 1107 any new objects being applied to the application are built into the application from the bottom up toward the root node of the application.
  • orphan control is established with respect to all of the components of the application that will be replaced or modified. Establishing orphan control involves identifying the components of the application that will be deleted, replaced, or modified, and establishing an orphan state of those components. The orphan state enables clients that are already queued for interaction with those components to traverse those components in a seamless manner.
  • step 1108 the state of each orphan established in the target voice application is continually checked for an opportunity to purge the orphan and allow a new object to take over that position in the dialog.
  • step 1109 it is decided whether those orphans checked have any callers interacting with them.
  • step 1110 if an orphan has callers interacting, the process reverts back to step 1108 for that orphan. All established orphans might, in one embodiment, be monitored simultaneously.
  • step 1108 if an orphan does not have calls interacting then at step 1109 that orphan may be purged if the new component associated therewith is already in place to take over from the orphan as a result of step 1107 .
  • a change is implemented only when a last maintained orphan of a tree is free of calls. Then the next orphan up is continually monitored in step 1108 until it is free of calls. In one embodiment however, if a change-order is only to modify certain content or style of one or more voice prompts of an application but does not change the intent or direction of the interaction flow with respect to caller position, then any orphan in the tree may be purged at step 1110 when it is not in a current interaction state. At step 1110 , a new object associated with an orphan immediately takes over when an orphan is purged. If an orphan has no replacement node it is simply purged when it is not currently in use.
  • the code portion of the new configuration provides all of the required linking functionality for establishing transient or temporary linking orders from prompt to prompt in a dialog. Therefore, an orphan that is still in use, for example, may be temporarily linked to a new node added further down the dialog tree. When that orphan is purged, a new object (if in place) takes over the responsibilities of caller interaction and linking to further objects.
  • the system reports status of task implementation.
  • files are actually swapped from cache to permanent storage during configuration.
  • a new component may not be inserted into the voice application until the final orphan being maintained in the tree is cleared of callers for a sufficient amount of time to make the change over and load the actual file or files representing the new object.
  • the next orphan above a newly inserted object may be automatically linked to the new component so that existing callers interacting with that orphan can seamlessly traverse to the new component in the application enabling lower orphan nodes to be purged. This process may evolve up tree of the voice application until all of the new objects are implemented and all orphans are purged.
  • new objects are installed immediately after orphans are established at step 1106 .
  • the new objects are installed side-by-side with the established orphans accept in the case where an orphan is deleted with no modification or replacement plan.
  • the new components are selected to immediately take over during a lull in interaction when there are currently no callers interacting with that portion of the tree.
  • New objects may also be added that do not replace or conflict with any existing files of a voice application. In this case no orphan control is required. Code and linking instruction in a new configuration is applied to the old configuration in the same manner as voice file prompts.
  • transitory links are established in a new configuration for the purpose of maintaining application dialog flow while new objects are installed. For example, 2 links, one to an orphan and one to the new component may be provided to an existing component that will be affected. If an orphan has current callers but the node below it has none, the orphan can automatically link to the new object even though it is still being used.
  • step 1107 installing new objects may include additional subroutines involving moving the objects from cache memory to permanent or semi-permanent storage. If the physical voice files and code replacements are preloaded into a database and then accessed during the configuration implementation, then step 1107 may proceed regardless of orphan status, however the new components are activated only according to orphan status.
  • the method and apparatus of the present invention can be implemented within or on a local area network, or from a remote point of access to a wide area network including the Internet network without departing from the spirit and scope of the present invention.
  • the software of the present invention can be adapted to any type of voice portal that users may interact with and that plays voice files according to a pre-determined order.
  • the method and apparatus of the present invention in light of many possible embodiments, some of which are described herein should be afforded the broadest possible scope under examination. The spirit and scope of the present invention is limited only by the following claims.

Abstract

A system for configuring and implementing changes to a voice application system has a first software component and host node for configuring one or more changes; a second software component and host node for receiving and implementing the configured change or changes; and a data network connecting the host nodes. In a preferred embodiment, a pre-configured change-order resulting from the first software component and host node is deployed after pre-configuration, deployment and execution thereof requiring only one action.

Description

    CROSS-REFERENCE TO RELATED DOCUMENTS
  • This application is a Continuation of co-pending U.S. patent application Ser. No. 10/861,078, filed Jun. 04, 2004, the disclosure of which is incorporated by reference herein.
  • That application claims priority to Provisional Application Ser. No. 60/558,921, filed on Apr. 02, 2004, and is a Continuation In Part (CIP) of co-pending U.S. patent application Ser. No. 10/835,444, filed Apr. 28, 2004, the disclosures of which are also incorporated by reference herein. U.S. patent application Ser. No. 10/835,444 claims priority to Provisional Application Ser. No. 60/532,271, filed on Dec. 22, 2003, the disclosure of which is also incorporated by reference herein.
  • TECHNICAL FIELD
  • The present invention is in the area of voice application software systems and pertains particularly to systems for developing and managing voice files linked for service to a voice application deployment system.
  • BACKGROUND
  • A speech application is one of the most challenging applications to develop, deploy and maintain in a communications environment. Expertise required for developing and deploying a viable VXML application, for example, includes expertise in computer telephony integration (CTI) hardware and software or a data network telephony (DNT) equivalent, voice recognition software, text-to-speech software, and speech application logic.
  • With the relatively recent advent of voice extensive markup language (VXML) the expertise require to develop a speech solution has been reduced somewhat. VXML is a language that enables a software developer to focus on the application logic of the voice application without being required to configure underlying telephony components. Typically, the developed voice application is run on a VXML interpreter that resides on and executes on the associated telephony system to deliver the solution.
  • Voice prompting systems in use to day range from a simple interactive voice response system for telephony to the more state-of-art VXML application system known to the inventors. Anywhere a customer telephony interface may be employed, there may also be a voice interaction system in place to interact with callers in real time. Data network telephony (DNT) equivalents of voice delivery systems also exist like VoIP portals and the like.
  • Often in both VXML compliant and non-VXML systems, such as computer telephony integrated (CTI) IVRs, voice messaging services and the like, voice prompts are sometimes prerecorded in a studio setting for a number of differing business scenarios and uploaded to the enterprise system server architecture for access and deployment during actual interaction with clients. Pre-recording voice prompts instead of dynamically creating them through software and voice synthesis methods is many times performed when better sound quality, different languages, different voice types, or a combination of the above are desired for the presentation logic of a particular system.
  • In very large enterprise architectures there may be many thousands of prerecorded voice prompts stored for use by a given voice application. Some of these may not be stored in a same centralized location. One with general knowledge of voice file management will attest that managing such a large volume of voice prompts can be very complicated. For example, in prior-art systems management of voice prompts includes recording the prompts, managing identification of those prompts and manually referencing the required prompts in the application code used in developing the application logic for deployment of those prompts to a client interfacing system. There is much room for error in code referencing and actual development, recording, and sorting batches of voice files can be error prone and time consuming as well.
  • The inventor knows of a software interface for managing audio resources used in one or more voice applications. The software interface includes a first portion for mapping the audio resources from storage to use-case positions in the one or more voice applications, a portion for accessing the audio resources according to the mapping information and for performing modifications a portion for creating new audio resources; and a portion for replication of modifications across distributed facilities. In a preferred application a developer can modify or replace existing audio resources and replicate links to the application code of the applications that use them.
  • VXML-compliant and other types of voice systems may frequently need to be modified or updated, sometimes multiple times per day due to fast-paced business environments, rapidly evolving business models, special temporary product promotions, sales discounts and so on. For example, if a product line goes obsolete, existing voice prompts related to that product line that are operational in a deployed voice application may need to be modified, replaced or simply deleted. Moreover, configuration settings of a voice application interaction system may also need to be updated or modified from time to time due to addition of new hardware, software, and so on.
  • The software application mentioned above as known to the inventor for managing audio resources enables frequent modifications of existing voice applications in a much improved and efficient manner than in the current art. However, when changing over from an existing configuration to a new configuration the running voice application is typically suspended from service while the changes are implemented. Shutting down service for even a temporary period can result in monetary loss that can be significant depending on the amount of time the system will be shut down. In some cases a backup system may be deployed while the primary system is being reconfigured. However this approach requires more resources that would be required to run one application.
  • What is clearly needed is a software routine or application for facilitating a one-click or single-action deployment and implementation of voice application system changes.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a logical overview of a voice interaction server and voice prompt data store according to prior-art.
  • FIG. 2 is a block diagram illustrating voice prompt development and linking to a voice prompt application according to prior art.
  • FIG. 3 is a block diagram illustrating a voice prompt development and management system according to an embodiment of the present invention.
  • FIG. 4 illustrates an interactive screen for a voice application resource management application according to an embodiment of the present invention.
  • FIG. 5 illustrates an interactive screen having audio resource details and dependencies according to an embodiment of the present invention.
  • FIG. 6 illustrates an interactive screen for an audio resource manager illustrating further details and options for editing and management according to an embodiment of the present invention.
  • FIG. 7 is a process flow diagram illustrating steps for editing or replacing an existing audio resource and replicating the resource to distributed storage facilities.
  • FIG. 8 is architectural overview of a communications network wherein automated voice application system configuration is practiced according to an embodiment of the present invention.
  • FIG. 9 is an exemplary screenshot illustrating application of modifications to a voice dialog according to an embodiment of the present invention.
  • FIG. 10 is a block diagram illustrating components of an automated voice application configuration application according to an embodiment of the present invention.
  • FIG. 11 is a process flow chart illustrating steps for receiving and implementing a change-order according to an embodiment of the present invention.
  • DETAILED DESCRIPTION
  • The inventor provides a system for managing voice prompts in a voice application system. Detail about methods, apparatus and the system as a whole are described in enabling detail below.
  • FIG. 1 is a logical overview of a voice interaction server and voice prompt data store according to prior art. FIG. 2 is a block diagram illustrating voice prompt development and linking to a voice prompt application according to prior art. A voice application system 100 includes a developer 101, a voice file storage medium 102, a voice portal (telephony, IVR) 103, and one of possibly hundreds or thousands of receiving devices 106.
  • Device 106 may be a LAN-line telephone, a cellular wireless, or any other communication device that supports voice and text communication over a network. In this example, device 106 is a plane old telephone service (POTS) telephone.
  • Device 106 has access through a typical telephone service network, represented herein by a voice link 110, to a voice system 103, which in this example is a standard telephony IVR system. IVR system 103 is the customer access point for callers (device 106) to any enterprise hosting or leasing the system.
  • IVR 103 has a database/resource adapter 109 for enabling access to off-system data. IVR also has voice applications 108 accessible therein and adapted to provide customer interaction and call flow management. Applications 108 include the capabilities of prompting a customer, taking input from a customer and playing prompts back to the customer depending on the input received.
  • Telephony hardware and software 107 includes the hardware and software that may be necessary for customer connection and management of call control protocols. IVR 103 may be a telephony switch enhanced as a customer interface by applications 108. Voice prompts executed within system 103 may include only prerecorded prompts. A DNT equivalent may use both prerecorded prompts and XML-based scripts that are interpreted by a text-to-speech engine and played using a sampled voice.
  • IVR system 103 has access to a voice file data store 102 via a data link 104, which may be a high-speed fiber optics link or another suitable data carrier many of which are known and available. Data store 102 is adapted to contain prerecorded voice files, sometimes referred to as prompts. Prompts are maintained, in this example, in a section 113 of data store 102 adapted for the purpose of storing them. A voice file index 112 is illustrated and provides a means for searching store section 113 to access files for transmission over link 104 to IVR system 103 to be played by one of applications 108 during interaction with a client.
  • In this case IVR system 102 is a distributed system such as to a telephony switch location in a public switched telephone network (PSTN) and therefore is not equipped to store many voice files, which take up considerable storage space if they are high quality recordings.
  • Data store 111 has a developer/enterprise interface 111 for enabling developers such as developer 101 access for revising existing voice files and storing new and deleting old voice files from the data store. Developer 101 may create voice applications and link stored voice files to the application code for each voice application created and deployed. Typically, the voice files themselves are created in a separate studio from script provided by the developer.
  • As was described with reference to the background section, for a large enterprise there may be many thousands of individual voice prompts, many of which are linked together in segmented prompts or prompts that are played in a voice application wherein the prompts contain more than one separate voice file. Manually linking the original files to the application code when creating the application provides enormous room for human error. Although the applications are typically tested before deployment, errors may still get through causing monetary loss at the point of customer interface.
  • Another point of human management is between the studio and the developer. The studio has to manage the files and present them to the developer in a fashion that the developer can manipulate in an organized fashion. As the number of individual prerecorded files increases, so does the complexity of managing those prerecorded files.
  • Referring now to FIG. 2, developer 101 engages in voice application development activity 201. Typically voice files are recorded from script. Therefore, for a particular application developer 101 creates enterprise scripts 202 and sends them out to a studio (200) to be recorded. An operator within studio 200 receives scripts 202 and creates recorded voice files 203. Typically, the files are single segments, some of which may be strategically linked together in a voice application to play as a single voice prompt to a client as part of a dialog executed from the point of IVR 103, for example.
  • The enterprise must insure that voice files 203 are all current and correct and that the parent application has all of the appropriate linking in the appropriate junctions so that the files may be called up correctly during execution. Developer 101 uploads files 203 when complete to data store 102 and the related application may also be uploaded to data store 102. When a specific application needs to be run at a customer interface, it may be distributed without the voice files to the point of interface, in this case IVR 103. There may be many separate applications or sub-dialogs that use the same individual voice files. Often there will be many instances of the same voice file stored in data store 102 but linked to separate applications that use the same prompt in some sequence.
  • FIG. 3 is an expanded view of IVR 103 of FIG. 2 illustrating a main dialog and sub-dialogs of a voice application according to prior art. In many systems, a main dialog 300 includes a static interactive menu 301 that is executed as part of the application logic for every client that calls in. During playing of menu 300, a client may provide input 302, typically in the form of voice for systems equipped with voice recognition technology. A system response 303 is played according to input 302.
  • System response 303 may include as options, sub-dialogs 304(a-n). Sub-dialogs 304(a-n) may link any number of prompts, or voice files 305(a-n) illustrated logically herein for each illustrated sub-dialog. In this case prompt 305 b is used in sub-dialog 304 a and in sub-dialog 304 b. Prompt 305 c is used in all three sub-dialogs illustrated. Prompt 305 a is used in sub-dialog 304 b and in sub-dialog 304 b. Prompts are created at the time of application creation and deployment. Therefore prompts 305 b, c, and j are stored in separate versions and locations for each voice application.
  • FIG. 4 illustrates an interactive screen 400 for a voice application resource management application according to an embodiment of the present invention. Screen 400 is a GUI portion of a software application that enables a developer to create and manage resources used in voice applications. Resources include both audio resources and application scripts that may be voice synthesized. For the purpose of this example, the inventor focuses on management of audio resources, which in this case, include voice file or prompt management in the context of one or more voice file applications.
  • Screen 400 takes the form of a Web browser type interface and can be used to access remote resources over a local area network (LAN), wide area network (WAN), or a metropolitan area network (MAN). In this example, a developer operating through screen 400 is accessing a local Intranet.
  • Screen 400 has a toolbar link 403 that is labeled workspace. Link 403 is adapted to open, upon invocation, a second window or changes the primary window to provide an area for working and audio management and creation tools for creating and working with audio files and transcripts or scripts.
  • Screen 400 has a toolbar link 404 that is labeled application. Link 404 is adapted to open, upon invocation, a second window or changes the primary window to provide an area for displaying and working with voice application code and provides audio resource linking capability. Screen 400 also has a toolbar link for enabling an administration view of all activity.
  • Screen 400 has additional toolbar links 406 adapted for navigating to different windows generally defined by label. Reading from left to right in toolbar options 406, there is Audio, Grammar, Data Adapter, and Thesaurus. The option Audio enables a user to view all audio-related resources. The option Grammar enables a user to view all grammar-related resources. The option Data Adapter enables a user to view all of the available adapters used with data sources, including adapters that might exist between disparate data formats. The option Thesaurus is self-descriptive.
  • In this example, a developer has accessed the audio resource view, which provides in window 409 an interactive data list 411 of existing audio resources currently available in the system. List 411 is divided into two columns a column 408 labeled “name” and a column 410 labeled “transcript”. In this example there are three illustrated audio prompts reading from top to bottom from list 411 column 408 they are “howmuch”, “mainmenu”, and “yourbalance”. An audio speaker icon next to each list item indicates the item is an audio resource and enable a developer to. Each audio resource is associated with the appropriate transcript of the resource as illustrated in column 410. Reading from top to bottom in column 410 for the audio resource “howmuch” the transcript is “How much do you wish to transfer?”. For “mainmenu”, the transcript is longer, therefore it in not reproduced in the illustration but may be assumed to be provided in full text. A scroll function may be provided to scroll a long transcript associated with an audio resource. For the audio resource “yourbalance”, the transcript is “Your balance is [ ]. The brackets enclose a variable used in a voice system prompt response to a client input interpreted by the system.
  • In one embodiment there may be additional options for viewing list 411, for example, separate views of directory 411 may be provided in different languages. In one embodiment, separate views of directory 411 may be provided for the same resources recorded using different voice talents. In the case of voice files that are contextually the same, but are recorded using different voice talents and or languages, those files may be stored together and versioned according to language and talent.
  • Window 409 can be scrollable to reach any audio resources not viewable in the immediate screen area. Likewise, in some embodiments a left-side navigation window may be provided that contains both audio resource and grammar resource indexes 401 and 402 respectively to enable quick navigation through the lists. A resource search function 411 is also provided in this example to enable keyword searching of audio and grammar resources.
  • Screen 400 has operational connectivity to a data store or stores used to where house the audio and grammar resources and, in some cases, the complete voice applications. Management actions initiated through the interface are applied automatically to the resources and voice applications.
  • A set of icons 407 defines additional interactive options for initiating immediate actions or views. For example, accounting from left to right a first icon enables creation of a new audio resource from a written script. Invocation of this icon brings up audio recording and editing tools that can be used to create new audio voice files and that can be used to edit or version existing audio voice files. A second icon is a recycle bin for deleting audio resources. A third icon in grouping 407 enables an audio resource to be copied. A fourth icon in grouping 407 enables a developer to view a dependency tree illustrating if where and when the audio file is used in one or more voice dialogs. The remaining two icons are upload and download icons enabling the movement of audio resources from local to remote and from remote to local storage devices.
  • In one embodiment of the present invention, the functions of creating voice files and linking them to voice applications can be coordinated through interface 400 by enabling an author of voice files password protected local or remote access for downloading enterprise scripts and for uploading new voice files to the enterprise voice file database. By marking audio resources in list 410 and invoking the icon 407 adapted to view audio resource dependencies, an operator calls up a next screen illustrating more detail about the resources and further options for editing and management as will be described below.
  • Screen 400, in this example, has and audio index display area 401 and a grammar display index area 402 strategically located in a left scrollable sub-window of screen 400. As detailed information is viewed for a resource in window 409, the same resource may be highlighted in the associated index 401 or 402 depending on the type of resource listed.
  • FIG. 5 is illustrates an interactive screen 500 showing audio resource details and dependencies according to an embodiment of the present invention. Screen 500 has a scrollable main window 501 that is adapted to display further details about audio resources previously selected for view. Previous options 406 remain displayed in screen 500. In this example each resource selected in screen 400 is displayed in list form. In this view audio resource 504 has a resource name “howmuch”. The resource 504 is categorized according to Dialog, Dialog type, and where the resource is used in existing voice applications. In the case of resource 504, the dialog reference is “How Much”, the resource type is a dialog, and the resource is used in a specified dialog prompt. Only one dependency is listed for audio resource 504, however all dependencies (if more than one) will be listed.
  • Resource 505, “mainmenu” has dependency to two main menus associated with dialogs. In the first listing the resource is used in a standard prompt used in the first listed dialog of the first listed main menu. In the second row it is illustrated that the same audio resource also is used in a nomatch prompt used in a specified dialog associated with the second listed main menu. For the purpose of this specification a nomatch prompt is one where the system does not have to match any data provided in a response to the prompt. A noinput prompt is one where no input is solicited by the prompt. It is noted herein that for a general application prompt definitions may vary widely according to voice application protocols and constructs used. The dependencies listed for resource 505 may be associated with entirely different voice applications used by the same enterprise. They may also reflect dependency of the resource to two separate menus and dialogs of a same voice application.
  • No specific ID information is illustrated in this example, but may be assumed to be present. For example, there may be rows and columns added for displaying a URL or URI path to the instance of the resource identified. Project Name, Project ID, Project Date, Recording Status (new vs. recorded), Voice Talent, and Audio Format are just some of the detailed information that may be made available in window 501. There may be a row or column added for provision of a general description of the resource including size, file format type, general content, and so on.
  • Resource 506, “yourbalance” is listed with no dependencies found for the resource. This may be because it is a newly uploaded resource that has not yet been linked to voice application code. It may be that it is a discarded resource that is still physically maintained in a database for possible future use. The lack of information tells the operator that the resource is currently not being used anywhere in the system.
  • Screen 500, in this example, has audio index display area 401 and a grammar display index area 402 strategically located in a left scrollable sub-window of screen 500 as described with reference to screen 400 of FIG. 4 above. As detailed information is viewed for a resource in window 501, the same resource may be highlighted in the associated index 401 or 402 depending on the type of resource listed.
  • FIG. 6 illustrates an interactive screen 600 of an audio resource manager illustrating further details and options for editing and management according to an embodiment of the present invention. Screen 600 enables a developer to edit existing voice files and to create new voice files. A dialog tree window 602 is provided and is adapted to list all of the existing prompts and voice files linked to dialogs in voice applications. The information is, in a preferred embodiment, navigable using a convenient directory and file system format. Any voice prompt or audio resource displayed in the main window 601 is highlighted in the tree of window 602.
  • In one embodiment of the present invention from screen 500 described above, a developer can download a batch of audio resources (files) from a studio remotely, or from local storage and can link those into an existing dialog, or can create a new dialog using the new files. The process, in a preferred embodiment, leverages an existing database program such as MS Excel™ for versioning and keeping track of voice prompts dialogs, sub-dialogs, and other options executed during voice interaction.
  • In one embodiment of the present invention a developer can navigate using the mapping feature through all of the voice application dialogs referencing any selected voice files. In a variation of this embodiment the dialogs can be presented in descending or ascending orders according to some criteria specified like date, number of use positions, or some other hierarchical specification. In still another embodiment, a developer accessing an audio resource may also have access to any associated reference files like coaching notes, contextual notes, voice talent preferences, language preferences, and pronunciation nuances for different regions.
  • In a preferred embodiment, using the software of the present invention multiple links do not have to be created to replace an audio resource used in multiple dialog prompts of one or more voice applications. For example, after modifying a single voice file, one click may cause the link to the stored resource to be updated across all instances of the file in all existing applications. In another embodiment where multiple storage sites are used, replication may be ordered such that the modified file is automatically replicated to all of the appropriate storage sites for local access. In this case, the resource linking is updated to each voice application using the file according to the replication location for that application.
  • Screen 600 illustrates a prompt 604 being developed or modified. The prompt in this example is named “Is that correct?” and has variable input fields of City and State. The prompt 604 combines audio files to recite “You said [City: State]:: If that is correct, say Yes: If in correct, Say No: The prompt may be used in more than one dialog in more than one voice application. The prompt may incorporate more than one individual prerecorded voice file.
  • A window 605 contains segment information associated with the prompt “Is that correct?” such as the variable City and State and the optional transcripts (actual transcripts of voice files). New voice files and transcripts describing new cities and states may be added and automatically liked to all of the appropriate prompt segments used in all dialogs and applications.
  • Typically, audio voice files of a same content definition, but prerecorded in one or more different languages and/or voice talents will be stored as separate versions of the file. However, automated voice translation utilities can be used to translate an English voice file into a Spanish voice file, for example, on the fly as the file is being accessed and utilized in an application. Therefore, in a more advanced embodiment multiple physical prerecorded voice files do not have to be maintained.
  • Screen 600 has a set of options 603 for viewing creating or editing prompts, rules, nomatch prompts, and no-input prompts. Options for help, viewing processor details, help with grammar, and properties are also provided within option set 603. Workspace provides input screen or windows for adding new material and changes. The workspace windows can be in the form of an excel worksheet as previously described.
  • In one embodiment of the present invention linking voice files to prompts in application can be managed across multiple servers in a distributed network environment. Voice files, associated transcripts, prompt positions, dialog positions, and application associations are all automatically applied for the editor eliminating prior-art practice of re-linking the new resources in the application code. Other options not illustrated in this example may also be provided without departing from the spirit and scope of the present invention. For example, when a voice file used in several places has been modified, the editor may not want the exact version to be automatically placed in all use instances. In this case, the previous file is retained and the editor simply calls up a list of the use positions and selects only the positions that the new file applies to. The system then applies the new linking for only the selected prompts and dialogs. The old file retains the linking to the appropriate instances where no modification was required.
  • In another embodiment, voice file replication across distributed storage systems is automated for multiple distributed IVR systems or VXML portals. For example, if a developer makes changes to voice files in one storage facility and links those changes to all known instances of their use at other client access points, which may be widely distributed, then the distributed instances may automatically order replication of the appropriate audio resources from the first storage facility to all of the other required storage areas. Therefore, for voice applications that are maintained at local client-access facilities of a large enterprise that rely on local storage of prerecorded files can, after receiving notification of voice file linking to a new file or files can execute and order to retrieve those files from the original storage location and deposit them into their local stores for immediate access. The linking then is used as a road map to insure that all distributed sites using the same applications have access to all of the required files. In this embodiment audio resource editing can be performed at any network address wherein the changes can be automatically applied to all distributed facilities over a WAN.
  • FIG. 7 is a process flow diagram 700 illustrating steps for editing or replacing an existing audio resource and replicating the resource to distributed storage facilities. At step 701, the developer selects an audio resource for edit or replacement. The selection can be based on a search action for a specific audio resource or from navigation through a voice application dialog menu tree.
  • At step 702 all dialogs that reference the selected audio resource are displayed. At step 703, the developer may select the dialogs that will use the edited or replacement resource by marking or highlighting those listed dialogs. In one embodiment all dialogs may be selected. The exact number of dialogs selected will depend on the enterprise purpose of the edit or replacement.
  • At step 704, the developer edits and tests the new resource, or creates an entirely new replacement resource. At step 705, the developer saves the final tested version of the resource. At step 706, the version saved is automatically replicated to the appropriate storage locations referenced by the dialogs selected in step 703.
  • In this exemplary process, steps 702, and step 706 are automated results of the previous actions performed.
  • The methods and apparatus of the present invention can be applied on a local network using a central or distributed storage system as well as over a WAN using distributed or central storage. Management can be performed locally or remotely, such as by logging onto the Internet or an Intranet to access the software using password protection and/or other authentication procedures.
  • The methods and apparatus of the present invention greatly enhance and streamline voice application development and deployment and according to the embodiments described, can be applied over a variety of different network architectures including DNT and POTS implementations.
  • One-Touch System Configuration Routine
  • According to one aspect of the present invention a software routine is provided that is capable of receiving a configuration package and of implementing the package at a point of voice interaction in order to effect system changes and voice application changes without suspending a system or application that is running and in the process of interaction with clients.
  • FIG. 8 is architectural overview of a communications network 800 wherein automated voice application system configuration is practiced according to an embodiment of the present invention. Communications network 800 encompasses a wide-area-network (WAN) 801, a public-switched-telephone-network (PSTN) 802, and a communications host illustrated herein as an enterprise 803.
  • Enterprise 803 may be any type of enterprise that provides services to clients, which are accessible to a call-in center or department. Enterprise 803, in this example, maintains voice interaction access points to voice services. Enterprise 803 may be assumed to contain a communications-center type environment wherein service agents interact with clients calling into the enterprise.
  • Enterprise 803 has a local-area-network (LAN) 820 provided therein and adapted for supporting a plurality of agent-operated workstations for communication and data sharing. LAN 820 has communications access to WAN 801 and to PSTN 802. A central telephony switch (CS) 821 is provided within enterprise 803 and is adapted to receive calls routed thereto from PSTN 802 via a telephony trunk branch 817 from a local switch in the network illustrated herein as switch (LS) 804. LS 804 may be a private-branch type of exchange (PBX), and automated-call-distributor (ACD), or any other type of telephone switch capable of running calls.
  • CS 821 has an interactive voice system peripheral (VS) 822 connected thereto by a CTI link. VS 822 also has connection to LAN 820. VS 822 is adapted to interact with callers routed CS 821 according to voice application dialogs therein. VS 822 may be an IVR system or a voice recognition system (VRS) without departing from the spirit and scope of the present invention. VS 822 is a point of deployment for voice applications used for client interaction. In this example, incoming calls routed to CS 821 from LS 800 from within PSTN 802 are illustrated as calls 805 incoming into LS 804 from anywhere within PSTN 805.
  • Enterprise 803 has a voice application server (VAS) 824 provided therein and connected to LAN 820. VAS 824 is adapted for storing and serving voice applications created by an administrator (ADMN) 823 represented herein by a computer icon also shown connected to LAN 820. Administrator 823 uses a client software application (AS) 825 to create voice applications and manage voice files, voice prompts, and voice dialogs of those applications.
  • Once applications are created they may be deployed by VAS 824 to VS 822 for immediate service. In one embodiment of the present invention, system 822 stores voice applications locally (storage not shown). In another embodiment of the present invention system 822 retrieves voice applications from system 824 over LAN 820 when those applications are required in interaction with clients. AS 825 installed on workstation 823 is analogous to an application described further above respect to screenshots 400, 500, and 600 of FIGS. 4, 5, and 6 respectively. One exception is that AS 825 is enhanced according to an embodiment of the present invention with a utility for enabling configuration and one touch deployment of voice application or system modification updates to voice applications or settings active at VS 822. In some embodiments of the present invention, updates created and deployed from workstation 823are applied to voice applications while those applications are active in interaction without a requirement for shutting down or suspending those applications from service.
  • Voice application server 824, in this embodiment, has connection to WAN 801 via a WAN access line 814. WAN 801 may be the well-known Internet, an Intranet, or a corporate WAN, among other possibilities. LAN access line 814 may be a 24/7 connection or a connection through a network service provider. WAN 801 has a network backbone 812 extending there through, which represents all of the lines, equipment, and access points making up the entire WAN as a whole.
  • Backbone 812 has a voice system (VS) 813 connected thereto, which represents a data-network-telephony (DNT) version of VS 822. System 813 uses voice applications to interact with clients accessing the system from anywhere in WAN 801 or any connected sub networks. It is noted herein, that networks 802 and 801 are bridged to gather for communication via a gateway 816. Gateway 816 is adapted translating telephony protocols into data network protocols and in reverse order enabling, for example, IP telephony callers to place calls to PSTN destinations, and PSTN telephony callers to place calls to WAN destinations. In one embodiment, gateway 816 may be an SS-7 Bell core system, or some other like system. Therefore, it is possible for PSTN callers to access voice interaction provided by system 813 and for WAN callers to access voice interaction provided by system 822.
  • A remote administrator is illustrated in this example as a remote administrator 818. Administrator 818 may be operating from a remote office, from a home, or from any physical location providing telephone and network-access services. A personal computer icon representing a workstation 819 further defines administrator 818. Workstation 819 is analogous in this embodiment to workstation 823 except that it is a remote workstation and not LAN-connected in this example.
  • Workstation 819 has a software application 825 a provided thereto, which is analogous to application 825 installed on workstation 823 within enterprise 803. Voice systems 822 and 813 have instances of a configuration order routine (COR), 826 for VS 822, and 826 a for VS 813 installed thereon. COR (826, 826 a) is adapted to except a configuration order package from AS 825 and/or AS 825 a respectively. COR (826, 826 a) excepts and implements configuration orders created by administrators 823 or 819 and automatically applies those configuration orders to their respective voice systems.
  • In a preferred embodiment of the present invention, administrator 823 utilizes application software 825 create necessary updates to existing voice applications including any required settings changes. Voice application server 824 contains the actual voice applications in this case, which may be served to voice system 822 when required. In one embodiment however, voice system 822 may store voice applications for immediate access. After making the required edits, administrator 823 may initiate a one-touch deployment action that causes a change-order to be implemented by change-order routine 826 running in VS 822. It is noted herein that a change-order for a voice application that is running may automatically extract and implement itself while the application is still running. A change-order may also be implemented to an application that is not currently running without departing from the spirit and scope of the present invention.
  • When VS 822 receives a change-order from administrator 823, application 826 executes and implements the change-order. In the case of a running application, there may be a plurality of callers queued for different dialog prompts or prompt sequences of the same application. In this case, COR 826 monitors the state of the running application and implements the changes so that they do not negatively affect caller interaction with the application. More detail about how this is accomplished is provided later in this specification.
  • Remote administrator 819 may also create and implement change-orders to applications running in voice system 822 from a remote location. For example, utilizing AS 825 a, administrator 819 may connect to ISP 809 through LS 804 via trunk 806 and trunk branch 808. ISP 809 may then connect administrator 819 to backbone 812, from where VS 824 is accessible via network line 814. Administrator 819 may therefore perform any of the types of edits or changes to applications running in VS 822 or to any settings of VS 822 that administrator 823 could configure for the same. Moreover, administrators 823 and 819 may generate updates for any voice applications running on voice system 813 connected to backbone 812 in WAN 801.
  • Calls 805 may represent PSTN callers accessing CS 821 through trunk 806 and trunk branch 817. Calls 805 may also include callers operating computers accessing VS 813 through ISP 809 via trunk branch 808 and network line 810, or through gateway 816 via trunk branch 807 and network line 815. Although the architecture in this example illustrates tethered access, callers 805 may also represent wireless users.
  • FIG. 9 is an exemplary interactive screen 900 illustrating application of modifications to a voice dialog according to an embodiment of the present invention. Screen 900 illustrates capability for creating a change-order or update to voice application dialog in this example. Screen 900 is a functional part of AS 825 or 825 a described above with reference to FIG. 8. Screenshot 900, in a preferred embodiment, stems from the same parent application hosting interactive screens 400, 500, and 600, described further above.
  • Interactive screen 900 contains a workspace 902, and a workspace 903. Space 902 contains a portion 904 of a dialog D-01 (logical representation only) illustrated in expanded view as a dialog 901, which is accessible from a dialog menu illustrated at far left of screen 900. A dialog search box is provided for locating any particular dialog that needs to be updated.
  • Within workspace 902, dialog portion 904 is illustrated in the form of an original configuration. In this example, a prompt 906 and a prompt 908 of dialog portion 904 will be affected by an update. Dialog portion 900 is illustrated within workspace 903 as an edited version 905. Workspace 903 is a new configuration workspace.
  • Prompt 906 in workspace 902 is to be replaced. In workspace 903, the affected prompt is illustrated as a dotted rectangle containing an R signifying replacement. In this example, prompt 906 is replaced with a prompt sequence 907. Sequence 907 contains three prompts labeled A signifying addition. Prompt 908 from workspace 902 is illustrated as a deleted prompt 909 in workspace 903 (dotted rectangle D).
  • The new configuration 905 can be “saved-to-file” by activating a save button 910, or can be saved and deployed by activating a deploy button 911. A reset button is also provided for resetting new configuration 905 to the form of the original the configuration 904. Interactive options for selecting prompts and for selecting attributes are provided for locating the appropriate new files link to the dialog. Each workspace 902 and 903 has a prompt-view option enabling an administrator to select any prompt in the tree and expand that prompt for play-back purposes or for viewing transcripts, author data, and so on.
  • When an original configuration has been updated to reflect a new configuration, selecting the deploy option 911 causes the update package to be deployed to the appropriate VS system (if stored therein) or to the VAS if the application is executed from such a server. The exact point of access for any voice system will depend on the purpose and design of the system. For example, referring back to FIG. 8, if a voice system and switch are provided locally within an enterprise, then the actual voice applications may be served to clients through the voice system, the application hosted on a separate machine, but called in to service when needed. In one embodiment, VS 824 distributes the voice applications to the respective interaction points or hosts, especially if the interaction host machine is remote.
  • FIG. 10 is a block diagram illustrating components of automated voice application configuration routine (826, 826 a) according to an embodiment of the present invention. Application 826 contains several components that enable automated configuration of updates or edits to voice applications that may be in the process of assisting clients.
  • Application 826 has a server port interface 1000 adapted to enable the application to detect when a change-order or update has arrived at the voice system. A host machine running application 826, in a preferred embodiment, will have a cache memory or data queue adapted to contain incoming updates to voice applications, some of which may be running when the updates have arrived.
  • Application 826 has a scheduler component provided therein and adapted to receive change-orders from a cache memory and schedule those change-orders for task loading. It is noted herein that a change-order may have its own schedule for task loading. In this case scheduler 1002 parses the schedule of the change-order and will not load the order until the correct time has arrived. Application 826 has a task loader 1003 provided therein and adapted to accept change-orders from scheduler 1002 for immediate implementation.
  • In one embodiment of the present invention, application 826 receives change-orders that include both instructions and the actual files required to complete the edits. In another embodiment of the present invention application 826 receives only the instructions, perhaps in the form of an object map or bitmap image, wherein the actual files are preloaded in identifiable fashion into a database containing the original files of the voice application or voice system settings. For updating voice applications, the actual implementation will depend on whether the voice files used to update the application are stored locally (within the VS) or are accessed from a separate machine such as a VAS.
  • Application 826 has a voice application (VA) locator 1004 provided therein, and adapted to find, in the case of voice application update, the correct application that will be updated. It is possible that the application being updated is not in use currently. It is also possible that the application being updated is currently in use. In either instance, VA locator 1004 is responsible for finding the location of the application and its base files.
  • VA locator 1004 has connection to a database or server base interface 1006 provided therein and adapted to enable VA locator 1004 to communicate externally from the host system or VS. Therefore, if a particular voice application is being stored on a voice application server separate from voice system that uses the interaction, the voice application locator running on the voice system can locate correct application on the external machine.
  • Application 826 has a voice application (VA) state monitor 1005 provided therein and adapted to monitor state of any voice application identified by VA locator 1004 that is currently running and serving clients during the time of update. State monitor 1005 has connection to a dialog controller interface 1009. A dialog controller is used by the voice system to execute a voice application. The dialog controller manages the caller access and dialog flow of any voice application in use by the system and therefore has state information regarding the number of clients interacting with the application and their positions in the dialog hierarchy.
  • Application 826 has a sub-task scheduler/execution module 1007 provided therein, and adapted to execute a change-order task according to instructions provided by VA state monitor 1005. Module 1007 contains an orphan controller 1008. Orphan controller 1008 is adapted to maintain a functioning state in a voice application of certain prompts or prompt sequences that are to be deleted or replaced with new files used by a new configuration.
  • It is important that the current client load using the voice application under modification is not inconvenienced in any way during the flow of the application and that clients traversing a new dialog will have the prompts in place so that the application does not crash. For this reason, orphans are maintained from the top down while changes to the application are built from the bottom up. In one embodiment of the present invention a new configuration is an object tree wherein the objects are prompts and prompt sequences. Similarly, the voice application that is to be modified has a similar object tree. The objects or nodes are links to the actual files that are applied in voice interaction. Likewise, there are objects or nodes in a voice application tree that represent functional code responsible for the direction of the application determined according to user response.
  • Module 1007 cooperates with VA state monitor 1005 to perform a change-order to a voice application using orphan controller 1008 to maintain functional orphans until all of the new objects are in place and callers are cleared from the orphan tree. In actual practice, the voice application being modified continues to function as a backup application while it is being modified. Replacement files and code modules associated with the change-order are, in a preferred embodiment, available in a same data store and memory partition that the original application files and code reside having been loaded therein either from cache or directly. In one embodiment, the files representing changes may be preloaded into the same storage hosting the old files such that as change-order is implemented by application 826 the change files are caused to take the place of the original files as required. The subtask scheduler portion of module 1007 works with VA state monitor 1005, which in turn has connection to the application dialog controller, which in turn has connection to the telephony hardware facilitating client connection to voice applications. Therefore module 1007 can apply changes to the application and maintain orphan state until all of the accessing callers are interacting with the new configuration in a seamless matter. At that point the orphans (old files and settings) may be purged from the system.
  • Application 826 has a task state/completion notification module 1010 provided therein and adapted to send notification of the completed task to the task author or administrator through server port interface 1000. Module 1010 also has connection to change-order cache interface 1001 for the purpose of purging the cache of any data associated with a task that has been completed successfully.
  • In one embodiment of the present invention, module 1010 may send, through interface 1000, an error notification or an advisory notification related to a change-order task that for some reason has not loaded successfully or that cannot be implemented efficiently. In the latter case, it may be that due to an unusually heavy call load using an existing application a change-order may be better scheduled during a time when there are not as many clients accessing the system. However, this is not required in practice the present invention as during change-order implementation, nodes are treated individually in terms of caller access and as long as the new changes are implemented from the bottom up callers may be transferred from an orphan, for example, to a new object in a dialog tree until such time that that orphan may be replaced or deleted and so on.
  • Application 826 may be provided as a software application or routine that takes instruction directly from the change-orders it receives. In one embodiment of the present invention application 826 may be provided to run on a piece of dedicated hardware as firmware, the hardware having connection to the voice system. There are much possible variant architecture that may be used without departing from the spirit and scope of the present invention.
  • FIG. 11 is a process flow chart 1100 illustrating steps or receiving and implementing a change the according to an embodiment of the present invention. At step 1101, a change-order is received by the system. In step 1101, the actual files of the change-order may be cached in a cache memory and the change-order instructions, which in one embodiment are of the form of an executable bitmap or object model, are loaded into a task loader analogous to loader 1003 of FIG. 10 for processing.
  • At step 1102, the system locates the voice application that is the target of the change-order. In one embodiment of the present invention, the target voice application may not be in current use. In this case, the changes may be implemented without concern for active state of the application interaction with clients. In another embodiment the target voice application may be currently in use with one or more of callers interacting with it. Assuming the latter case at step 1103, the system prepares for execution of the change implementation task. At step 1104, the current running state of the voice application is acquired. This information may include the total number of callers currently interacting with the application and their current positions of interaction with the application. Step 1104 is an ongoing step meaning that the system constantly receives current application state with respect to the number of callers and caller position in the dialog flow of the application.
  • At step 1105, execution of the change-order begins. At step 1106, any orphans in the old application are identified and maintained from the top or root node of the application down the hierarchy until they are idle or not in a current state of access from one or more clients. At step 1107, any new objects being applied to the application are built into the application from the bottom up toward the root node of the application. In step 1106, orphan control is established with respect to all of the components of the application that will be replaced or modified. Establishing orphan control involves identifying the components of the application that will be deleted, replaced, or modified, and establishing an orphan state of those components. The orphan state enables clients that are already queued for interaction with those components to traverse those components in a seamless manner.
  • At step 1108, the state of each orphan established in the target voice application is continually checked for an opportunity to purge the orphan and allow a new object to take over that position in the dialog. At step 1109, it is decided whether those orphans checked have any callers interacting with them. At step 1110, if an orphan has callers interacting, the process reverts back to step 1108 for that orphan. All established orphans might, in one embodiment, be monitored simultaneously. At step 1108, if an orphan does not have calls interacting then at step 1109 that orphan may be purged if the new component associated therewith is already in place to take over from the orphan as a result of step 1107.
  • In one embodiment of the present invention, a change is implemented only when a last maintained orphan of a tree is free of calls. Then the next orphan up is continually monitored in step 1108 until it is free of calls. In one embodiment however, if a change-order is only to modify certain content or style of one or more voice prompts of an application but does not change the intent or direction of the interaction flow with respect to caller position, then any orphan in the tree may be purged at step 1110 when it is not in a current interaction state. At step 1110, a new object associated with an orphan immediately takes over when an orphan is purged. If an orphan has no replacement node it is simply purged when it is not currently in use.
  • In a preferred embodiment of the present invention at steps 1106 and 1107, the code portion of the new configuration provides all of the required linking functionality for establishing transient or temporary linking orders from prompt to prompt in a dialog. Therefore, an orphan that is still in use, for example, may be temporarily linked to a new node added further down the dialog tree. When that orphan is purged, a new object (if in place) takes over the responsibilities of caller interaction and linking to further objects. At step 1111, the system reports status of task implementation.
  • In one embodiment of the present invention, files are actually swapped from cache to permanent storage during configuration. For example, a new component may not be inserted into the voice application until the final orphan being maintained in the tree is cleared of callers for a sufficient amount of time to make the change over and load the actual file or files representing the new object. The next orphan above a newly inserted object may be automatically linked to the new component so that existing callers interacting with that orphan can seamlessly traverse to the new component in the application enabling lower orphan nodes to be purged. This process may evolve up tree of the voice application until all of the new objects are implemented and all orphans are purged.
  • In a preferred application of the present invention, new objects are installed immediately after orphans are established at step 1106. In this embodiment, the new objects are installed side-by-side with the established orphans accept in the case where an orphan is deleted with no modification or replacement plan. In this case, the new components are selected to immediately take over during a lull in interaction when there are currently no callers interacting with that portion of the tree. New objects may also be added that do not replace or conflict with any existing files of a voice application. In this case no orphan control is required. Code and linking instruction in a new configuration is applied to the old configuration in the same manner as voice file prompts.
  • In one embodiment, transitory links are established in a new configuration for the purpose of maintaining application dialog flow while new objects are installed. For example, 2 links, one to an orphan and one to the new component may be provided to an existing component that will be affected. If an orphan has current callers but the node below it has none, the orphan can automatically link to the new object even though it is still being used.
  • One with skill in the art will recognize that the process order of flowchart 1100 may very according to the type of implementation. For example, if a change-order includes the physical voice files and code replacements and those are handled by the application, then at step 1107 installing new objects may include additional subroutines involving moving the objects from cache memory to permanent or semi-permanent storage. If the physical voice files and code replacements are preloaded into a database and then accessed during the configuration implementation, then step 1107 may proceed regardless of orphan status, however the new components are activated only according to orphan status.
  • The method and apparatus of the present invention can be implemented within or on a local area network, or from a remote point of access to a wide area network including the Internet network without departing from the spirit and scope of the present invention. The software of the present invention can be adapted to any type of voice portal that users may interact with and that plays voice files according to a pre-determined order. The method and apparatus of the present invention, in light of many possible embodiments, some of which are described herein should be afforded the broadest possible scope under examination. The spirit and scope of the present invention is limited only by the following claims.

Claims (16)

1. A method comprising:
selecting an original voice dialogue from a graphical library of dialogues, the graphical library of dialogues associated with voice applications and voice application servers where they are used;
editing the original voice dialogue to create an edited voice dialogue;
determining a number of voice systems and voice application servers using the original voice dialogue; and
deploying the edited voice dialogue to the voice systems and the voice application servers using the original voice dialogue, wherein the edited voice dialogue replaces the original version in the library.
2. The method of claim 1 wherein the edited voice dialogue includes at least one new voice prompt.
3. The method of claim 1 wherein the edited voice dialogue includes at least one new voice prompt sequence for a voice application.
4. The method of claim 1 wherein the edited voice dialogue includes at least one new core module.
5. The method of claim 1 wherein the edited voice dialogue is deployed to an interactive voice server via a data network.
6. The method of claim 1 further comprising displaying the original voice dialogue graphically as a connected series of prompts
7. An apparatus comprising:
a first workspace displaying an original voice dialogue;
a second workspace to edit the original voice dialogue into a edited voice dialogue;
a graphical library to maintain a plurality of voice dialogues, the graphical library to associate each of the plurality of voice dialogues with a voice application and a voice application server;
a counting device to determine a number of voice systems and voice application servers using the original voice dialogue;
a deployment module to deploy the edited voice dialogue to the voice systems and the voice application servers such that the edited voice dialog replaces the original voice dialogue.
8. The apparatus of claim 7 wherein the edited voice dialogue includes at least one new voice prompt.
9. The apparatus of claim 7 wherein the edited voice dialogue includes at least one new voice prompt sequence for a voice application.
10. The apparatus of claim 7 wherein the edited voice dialogue is deployed to an interactive voice server via a data network.
11. One or more computer readable media having stored thereon instructions that, when executed by a computing device, causes the computing device to:
select an original voice dialog from a library of dialogues, the library of dialogues associated with voice applications and voice application servers where they are used;
edit the original voice dialogue to create an edited voice dialogue;
determine a number of voice systems and voice application servers using the original voice dialogue; and
deploy the edited voice dialogue to the voice systems and the voice application servers using the original voice dialogue, wherein the edited voice dialogue replaces the original version in the library.
12. The computer readable media of claim 11 wherein the edited voice dialogue includes at least one new voice prompt.
13. The computer readable media of claim 11 wherein the edited voice dialogue includes at least one new voice prompt sequence for a voice application.
14. The computer readable media of claim 11 wherein the edited voice dialogue includes at least one new core module.
15. The computer readable media of claim 11 wherein the edited voice dialogue is deployed to an interactive voice server via a data network.
16. The computer readable media of claim 11 wherein the computing device further displays the original voice dialogue graphically as a connected series of prompts.
US11/728,954 2003-12-23 2007-03-27 Method for creating and deploying system changes in a voice application system Abandoned US20070219803A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US11/728,954 US20070219803A1 (en) 2003-12-23 2007-03-27 Method for creating and deploying system changes in a voice application system

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
US53227103P 2003-12-23 2003-12-23
US55892104P 2004-04-02 2004-04-02
US10/835,444 US7817784B2 (en) 2003-12-23 2004-04-28 System for managing voice files of a voice prompt server
US10/861,078 US7206391B2 (en) 2003-12-23 2004-06-04 Method for creating and deploying system changes in a voice application system
US11/728,954 US20070219803A1 (en) 2003-12-23 2007-03-27 Method for creating and deploying system changes in a voice application system

Related Parent Applications (2)

Application Number Title Priority Date Filing Date
US10/835,444 Continuation-In-Part US7817784B2 (en) 2003-12-23 2004-04-28 System for managing voice files of a voice prompt server
US10/861,078 Continuation US7206391B2 (en) 2003-11-17 2004-06-04 Method for creating and deploying system changes in a voice application system

Publications (1)

Publication Number Publication Date
US20070219803A1 true US20070219803A1 (en) 2007-09-20

Family

ID=35197648

Family Applications (2)

Application Number Title Priority Date Filing Date
US10/861,078 Active 2025-06-30 US7206391B2 (en) 2003-11-17 2004-06-04 Method for creating and deploying system changes in a voice application system
US11/728,954 Abandoned US20070219803A1 (en) 2003-12-23 2007-03-27 Method for creating and deploying system changes in a voice application system

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US10/861,078 Active 2025-06-30 US7206391B2 (en) 2003-11-17 2004-06-04 Method for creating and deploying system changes in a voice application system

Country Status (2)

Country Link
US (2) US7206391B2 (en)
WO (1) WO2005104472A2 (en)

Cited By (126)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7660719B1 (en) * 2004-08-19 2010-02-09 Bevocal Llc Configurable information collection system, method and computer program product utilizing speech recognition
US20100251247A1 (en) * 2009-03-27 2010-09-30 Ebay Inc. Change management automation tool
US20100280819A1 (en) * 2009-05-01 2010-11-04 Alpine Electronics, Inc. Dialog Design Apparatus and Method
US20110044437A1 (en) * 2004-04-28 2011-02-24 Apptera, Inc. Method and System for Presenting Dynamic Commercial Content to Clients Interacting with a Voice Extensible Markup Language system
US20110172994A1 (en) * 2010-01-13 2011-07-14 Apple Inc. Processing of voice inputs
US8024196B1 (en) * 2005-09-19 2011-09-20 Sap Ag Techniques for creating and translating voice applications
US20120081371A1 (en) * 2009-05-01 2012-04-05 Inci Ozkaragoz Dialog design tool and method
US8509403B2 (en) 2003-11-17 2013-08-13 Htc Corporation System for advertisement selection, placement and delivery
US8892446B2 (en) 2010-01-18 2014-11-18 Apple Inc. Service orchestration for intelligent automated assistant
US9262612B2 (en) 2011-03-21 2016-02-16 Apple Inc. Device access using voice authentication
US9300784B2 (en) 2013-06-13 2016-03-29 Apple Inc. System and method for emergency calls initiated by voice command
US9330720B2 (en) 2008-01-03 2016-05-03 Apple Inc. Methods and apparatus for altering audio output signals
US9338493B2 (en) 2014-06-30 2016-05-10 Apple Inc. Intelligent automated assistant for TV user interactions
US9368114B2 (en) 2013-03-14 2016-06-14 Apple Inc. Context-sensitive handling of interruptions
US9430463B2 (en) 2014-05-30 2016-08-30 Apple Inc. Exemplar-based natural language processing
US9483461B2 (en) 2012-03-06 2016-11-01 Apple Inc. Handling speech synthesis of content for multiple languages
US9495129B2 (en) 2012-06-29 2016-11-15 Apple Inc. Device, method, and user interface for voice-activated navigation and browsing of a document
US9502031B2 (en) 2014-05-27 2016-11-22 Apple Inc. Method for supporting dynamic grammars in WFST-based ASR
US9535906B2 (en) 2008-07-31 2017-01-03 Apple Inc. Mobile device having human language translation capability with positional feedback
US9576574B2 (en) 2012-09-10 2017-02-21 Apple Inc. Context-sensitive handling of interruptions by intelligent digital assistant
US9582608B2 (en) 2013-06-07 2017-02-28 Apple Inc. Unified ranking with entropy-weighted information for phrase-based semantic auto-completion
US9606986B2 (en) 2014-09-29 2017-03-28 Apple Inc. Integrated word N-gram and class M-gram language models
US9620104B2 (en) 2013-06-07 2017-04-11 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
US9620105B2 (en) 2014-05-15 2017-04-11 Apple Inc. Analyzing audio input for efficient speech and music recognition
US9626955B2 (en) 2008-04-05 2017-04-18 Apple Inc. Intelligent text-to-speech conversion
US9633004B2 (en) 2014-05-30 2017-04-25 Apple Inc. Better resolution when referencing to concepts
US9633674B2 (en) 2013-06-07 2017-04-25 Apple Inc. System and method for detecting errors in interactions with a voice-based digital assistant
US9633660B2 (en) 2010-02-25 2017-04-25 Apple Inc. User profiling for voice input processing
US9646609B2 (en) 2014-09-30 2017-05-09 Apple Inc. Caching apparatus for serving phonetic pronunciations
US9646614B2 (en) 2000-03-16 2017-05-09 Apple Inc. Fast, language-independent method for user authentication by voice
US9668121B2 (en) 2014-09-30 2017-05-30 Apple Inc. Social reminders
US9672211B1 (en) * 2015-04-07 2017-06-06 West Corporation Script unique prompts
US9697822B1 (en) 2013-03-15 2017-07-04 Apple Inc. System and method for updating an adaptive speech recognition model
US9697820B2 (en) 2015-09-24 2017-07-04 Apple Inc. Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks
US9711141B2 (en) 2014-12-09 2017-07-18 Apple Inc. Disambiguating heteronyms in speech synthesis
US9715875B2 (en) 2014-05-30 2017-07-25 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US9721566B2 (en) 2015-03-08 2017-08-01 Apple Inc. Competing devices responding to voice triggers
US9734193B2 (en) 2014-05-30 2017-08-15 Apple Inc. Determining domain salience ranking from ambiguous words in natural speech
US9760559B2 (en) 2014-05-30 2017-09-12 Apple Inc. Predictive text input
US9785630B2 (en) 2014-05-30 2017-10-10 Apple Inc. Text prediction using combined word N-gram and unigram language models
US9798393B2 (en) 2011-08-29 2017-10-24 Apple Inc. Text correction processing
US9818400B2 (en) 2014-09-11 2017-11-14 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US9842101B2 (en) 2014-05-30 2017-12-12 Apple Inc. Predictive conversion of language input
US9842105B2 (en) 2015-04-16 2017-12-12 Apple Inc. Parsimonious continuous-space phrase representations for natural language processing
US9858925B2 (en) 2009-06-05 2018-01-02 Apple Inc. Using context information to facilitate processing of commands in a virtual assistant
US9865280B2 (en) 2015-03-06 2018-01-09 Apple Inc. Structured dictation using intelligent automated assistants
US9886953B2 (en) 2015-03-08 2018-02-06 Apple Inc. Virtual assistant activation
US9886432B2 (en) 2014-09-30 2018-02-06 Apple Inc. Parsimonious handling of word inflection via categorical stem + suffix N-gram language models
US9899019B2 (en) 2015-03-18 2018-02-20 Apple Inc. Systems and methods for structured stem and suffix language models
US9922642B2 (en) 2013-03-15 2018-03-20 Apple Inc. Training an at least partial voice command system
US9934775B2 (en) 2016-05-26 2018-04-03 Apple Inc. Unit-selection text-to-speech synthesis based on predicted concatenation parameters
US9953088B2 (en) 2012-05-14 2018-04-24 Apple Inc. Crowd sourcing information to fulfill user requests
US9959870B2 (en) 2008-12-11 2018-05-01 Apple Inc. Speech recognition involving a mobile device
US9966065B2 (en) 2014-05-30 2018-05-08 Apple Inc. Multi-command single utterance input method
US9966068B2 (en) 2013-06-08 2018-05-08 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
US9971774B2 (en) 2012-09-19 2018-05-15 Apple Inc. Voice-based media searching
US9972304B2 (en) 2016-06-03 2018-05-15 Apple Inc. Privacy preserving distributed evaluation framework for embedded personalized systems
US10043516B2 (en) 2016-09-23 2018-08-07 Apple Inc. Intelligent automated assistant
US10049663B2 (en) 2016-06-08 2018-08-14 Apple, Inc. Intelligent automated assistant for media exploration
US10049668B2 (en) 2015-12-02 2018-08-14 Apple Inc. Applying neural network language models to weighted finite state transducers for automatic speech recognition
US10057736B2 (en) 2011-06-03 2018-08-21 Apple Inc. Active transport based notifications
US10067938B2 (en) 2016-06-10 2018-09-04 Apple Inc. Multilingual word prediction
US10074360B2 (en) 2014-09-30 2018-09-11 Apple Inc. Providing an indication of the suitability of speech recognition
US10079014B2 (en) 2012-06-08 2018-09-18 Apple Inc. Name recognition system
US10078631B2 (en) 2014-05-30 2018-09-18 Apple Inc. Entropy-guided text prediction using combined word and character n-gram language models
US10083688B2 (en) 2015-05-27 2018-09-25 Apple Inc. Device voice control for selecting a displayed affordance
US10089072B2 (en) 2016-06-11 2018-10-02 Apple Inc. Intelligent device arbitration and control
US10101822B2 (en) 2015-06-05 2018-10-16 Apple Inc. Language input correction
US10127220B2 (en) 2015-06-04 2018-11-13 Apple Inc. Language identification from short strings
US10127911B2 (en) 2014-09-30 2018-11-13 Apple Inc. Speaker identification and unsupervised speaker adaptation techniques
US10134385B2 (en) 2012-03-02 2018-11-20 Apple Inc. Systems and methods for name pronunciation
US10170123B2 (en) 2014-05-30 2019-01-01 Apple Inc. Intelligent assistant for home automation
US10176167B2 (en) 2013-06-09 2019-01-08 Apple Inc. System and method for inferring user intent from speech inputs
US10186254B2 (en) 2015-06-07 2019-01-22 Apple Inc. Context-based endpoint detection
US10185542B2 (en) 2013-06-09 2019-01-22 Apple Inc. Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant
US10192552B2 (en) 2016-06-10 2019-01-29 Apple Inc. Digital assistant providing whispered speech
US10199051B2 (en) 2013-02-07 2019-02-05 Apple Inc. Voice trigger for a digital assistant
US10223066B2 (en) 2015-12-23 2019-03-05 Apple Inc. Proactive assistance based on dialog communication between devices
US10241752B2 (en) 2011-09-30 2019-03-26 Apple Inc. Interface for a virtual digital assistant
US10241644B2 (en) 2011-06-03 2019-03-26 Apple Inc. Actionable reminder entries
US10249300B2 (en) 2016-06-06 2019-04-02 Apple Inc. Intelligent list reading
US10255907B2 (en) 2015-06-07 2019-04-09 Apple Inc. Automatic accent detection using acoustic models
US10269345B2 (en) 2016-06-11 2019-04-23 Apple Inc. Intelligent task discovery
US10276170B2 (en) 2010-01-18 2019-04-30 Apple Inc. Intelligent automated assistant
US10283110B2 (en) 2009-07-02 2019-05-07 Apple Inc. Methods and apparatuses for automatic speech recognition
US10289433B2 (en) 2014-05-30 2019-05-14 Apple Inc. Domain specific language for encoding assistant dialog
US10297253B2 (en) 2016-06-11 2019-05-21 Apple Inc. Application integration with a digital assistant
US10318871B2 (en) 2005-09-08 2019-06-11 Apple Inc. Method and apparatus for building an intelligent automated assistant
US10332518B2 (en) 2017-05-09 2019-06-25 Apple Inc. User interface for correcting recognition errors
US10356243B2 (en) 2015-06-05 2019-07-16 Apple Inc. Virtual assistant aided communication with 3rd party service in a communication session
US10354011B2 (en) 2016-06-09 2019-07-16 Apple Inc. Intelligent automated assistant in a home environment
US10366158B2 (en) 2015-09-29 2019-07-30 Apple Inc. Efficient word encoding for recurrent neural network language models
US10410637B2 (en) 2017-05-12 2019-09-10 Apple Inc. User-specific acoustic models
US10446143B2 (en) 2016-03-14 2019-10-15 Apple Inc. Identification of voice inputs providing credentials
US10446141B2 (en) 2014-08-28 2019-10-15 Apple Inc. Automatic speech recognition based on user feedback
US10482874B2 (en) 2017-05-15 2019-11-19 Apple Inc. Hierarchical belief states for digital assistants
US10490187B2 (en) 2016-06-10 2019-11-26 Apple Inc. Digital assistant providing automated status report
US10496753B2 (en) 2010-01-18 2019-12-03 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US10509862B2 (en) 2016-06-10 2019-12-17 Apple Inc. Dynamic phrase expansion of language input
US10521466B2 (en) 2016-06-11 2019-12-31 Apple Inc. Data driven natural language event detection and classification
US10553209B2 (en) 2010-01-18 2020-02-04 Apple Inc. Systems and methods for hands-free notification summaries
US10552013B2 (en) 2014-12-02 2020-02-04 Apple Inc. Data detection
US10567477B2 (en) 2015-03-08 2020-02-18 Apple Inc. Virtual assistant continuity
US10568032B2 (en) 2007-04-03 2020-02-18 Apple Inc. Method and system for operating a multi-function portable electronic device using voice-activation
US10593346B2 (en) 2016-12-22 2020-03-17 Apple Inc. Rank-reduced token representation for automatic speech recognition
US10592095B2 (en) 2014-05-23 2020-03-17 Apple Inc. Instantaneous speaking of content on touch devices
US10659851B2 (en) 2014-06-30 2020-05-19 Apple Inc. Real-time digital assistant knowledge updates
US10671428B2 (en) 2015-09-08 2020-06-02 Apple Inc. Distributed personal assistant
US10679605B2 (en) 2010-01-18 2020-06-09 Apple Inc. Hands-free list-reading by intelligent automated assistant
US10691473B2 (en) 2015-11-06 2020-06-23 Apple Inc. Intelligent automated assistant in a messaging environment
US10705794B2 (en) 2010-01-18 2020-07-07 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US10706373B2 (en) 2011-06-03 2020-07-07 Apple Inc. Performing actions associated with task items that represent tasks to perform
US10733993B2 (en) 2016-06-10 2020-08-04 Apple Inc. Intelligent digital assistant in a multi-tasking environment
US10747498B2 (en) 2015-09-08 2020-08-18 Apple Inc. Zero latency digital assistant
US10755703B2 (en) 2017-05-11 2020-08-25 Apple Inc. Offline personal assistant
US10762293B2 (en) 2010-12-22 2020-09-01 Apple Inc. Using parts-of-speech tagging and named entity recognition for spelling correction
US10789041B2 (en) 2014-09-12 2020-09-29 Apple Inc. Dynamic thresholds for always listening speech trigger
US10791216B2 (en) 2013-08-06 2020-09-29 Apple Inc. Auto-activating smart responses based on activities from remote devices
US10789945B2 (en) 2017-05-12 2020-09-29 Apple Inc. Low-latency intelligent automated assistant
US10791176B2 (en) 2017-05-12 2020-09-29 Apple Inc. Synchronization and task delegation of a digital assistant
US10810274B2 (en) 2017-05-15 2020-10-20 Apple Inc. Optimizing dialogue policy decisions for digital assistants using implicit feedback
US11010550B2 (en) 2015-09-29 2021-05-18 Apple Inc. Unified language modeling framework for word prediction, auto-completion and auto-correction
US11025565B2 (en) 2015-06-07 2021-06-01 Apple Inc. Personalized prediction of responses for instant messaging
US11217255B2 (en) 2017-05-16 2022-01-04 Apple Inc. Far-field extension for digital assistant services
US11281993B2 (en) 2016-12-05 2022-03-22 Apple Inc. Model and ensemble compression for metric learning
US11587559B2 (en) 2015-09-30 2023-02-21 Apple Inc. Intelligent device identification

Families Citing this family (52)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7406418B2 (en) * 2001-07-03 2008-07-29 Apptera, Inc. Method and apparatus for reducing data traffic in a voice XML application distribution system through cache optimization
US7609829B2 (en) * 2001-07-03 2009-10-27 Apptera, Inc. Multi-platform capable inference engine and universal grammar language adapter for intelligent voice application execution
US20030007609A1 (en) * 2001-07-03 2003-01-09 Yuen Michael S. Method and apparatus for development, deployment, and maintenance of a voice software application for distribution to one or more consumers
US7212615B2 (en) * 2002-05-31 2007-05-01 Scott Wolmuth Criteria based marketing for telephone directory assistance
US7206391B2 (en) * 2003-12-23 2007-04-17 Apptera Inc. Method for creating and deploying system changes in a voice application system
US20050234725A1 (en) * 2004-04-20 2005-10-20 International Business Machines Corporation Method and system for flexible usage of a graphical call flow builder
US7925512B2 (en) * 2004-05-19 2011-04-12 Nuance Communications, Inc. Method, system, and apparatus for a voice markup language interpreter and voice browser
EP1612977A3 (en) * 2004-07-01 2013-08-21 Yamaha Corporation Control device for controlling audio signal processing device
US9367846B2 (en) * 2004-11-29 2016-06-14 Jingle Networks, Inc. Telephone search supported by advertising based on past history of requests
US20060149553A1 (en) * 2005-01-05 2006-07-06 At&T Corp. System and method for using a library to interactively design natural language spoken dialog systems
US8185399B2 (en) 2005-01-05 2012-05-22 At&T Intellectual Property Ii, L.P. System and method of providing an automated data-collection in spoken dialog systems
US8478589B2 (en) 2005-01-05 2013-07-02 At&T Intellectual Property Ii, L.P. Library of existing spoken dialog data for use in generating new natural language spoken dialog systems
US9253301B2 (en) * 2005-09-01 2016-02-02 Xtone Networks, Inc. System and method for announcing and routing incoming telephone calls using a distributed voice application execution system architecture
US9426269B2 (en) * 2005-09-01 2016-08-23 Xtone Networks, Inc. System and method for performing certain actions based upon a dialed telephone number
WO2007028128A2 (en) 2005-09-01 2007-03-08 Vishal Dhawan Voice application network platform
US9456068B2 (en) 2005-09-01 2016-09-27 Xtone, Inc. System and method for connecting a user to business services
US20100158217A1 (en) 2005-09-01 2010-06-24 Vishal Dhawan System and method for placing telephone calls using a distributed voice application execution system architecture
US11153425B2 (en) * 2005-09-01 2021-10-19 Xtone, Inc. System and method for providing interactive services
US8964960B2 (en) * 2005-09-01 2015-02-24 Xtone Networks, Inc. System and method for interacting with a user via a variable volume and variable tone audio prompt
US20100166161A1 (en) * 2005-09-01 2010-07-01 Vishal Dhawan System and methods for providing voice messaging services
US9313307B2 (en) * 2005-09-01 2016-04-12 Xtone Networks, Inc. System and method for verifying the identity of a user by voiceprint analysis
US9799039B2 (en) * 2005-09-01 2017-10-24 Xtone, Inc. System and method for providing television programming recommendations and for automated tuning and recordation of television programs
US11102342B2 (en) 2005-09-01 2021-08-24 Xtone, Inc. System and method for displaying the history of a user's interaction with a voice application
US20070121873A1 (en) * 2005-11-18 2007-05-31 Medlin Jennifer P Methods, systems, and products for managing communications
US7773731B2 (en) * 2005-12-14 2010-08-10 At&T Intellectual Property I, L. P. Methods, systems, and products for dynamically-changing IVR architectures
US7627476B2 (en) * 2005-12-16 2009-12-01 International Business Machines Corporation Call flow modification based on user situation
US7577664B2 (en) 2005-12-16 2009-08-18 At&T Intellectual Property I, L.P. Methods, systems, and products for searching interactive menu prompting system architectures
US20070186165A1 (en) * 2006-02-07 2007-08-09 Pudding Ltd. Method And Apparatus For Electronically Providing Advertisements
US20080228479A1 (en) * 2006-02-24 2008-09-18 Viva Transcription Coporation Data transcription and management system and method
US7961856B2 (en) * 2006-03-17 2011-06-14 At&T Intellectual Property I, L. P. Methods, systems, and products for processing responses in prompting systems
US8050392B2 (en) * 2006-03-17 2011-11-01 At&T Intellectual Property I, L.P. Methods systems, and products for processing responses in prompting systems
US20070261027A1 (en) * 2006-05-08 2007-11-08 International Business Machines Corporation Method and system for automatically discovering and populating a palette of reusable dialog components
US7937687B2 (en) * 2006-09-01 2011-05-03 Verizon Patent And Licensing Inc. Generating voice extensible markup language (VXML) documents
US7747442B2 (en) * 2006-11-21 2010-06-29 Sap Ag Speech recognition application grammar modeling
US20080154590A1 (en) * 2006-12-22 2008-06-26 Sap Ag Automated speech recognition application testing
US20090100340A1 (en) * 2007-10-10 2009-04-16 Microsoft Corporation Associative interface for personalizing voice data access
US8635188B2 (en) * 2008-06-12 2014-01-21 Microsoft Corporation Techniques for extracting data from content databases
US8788460B2 (en) * 2008-06-12 2014-07-22 Microsoft Corporation Exploring attached and unattached content databases
US20140289184A1 (en) * 2009-09-09 2014-09-25 Sanjeev Kumar Biswas License structure representation for license management
US11054970B2 (en) * 2015-10-15 2021-07-06 Interactive Intelligence Group, Inc. System and method for multi-language communication sequencing
US11043206B2 (en) 2017-05-18 2021-06-22 Aiqudo, Inc. Systems and methods for crowdsourced actions and commands
US11056105B2 (en) * 2017-05-18 2021-07-06 Aiqudo, Inc Talk back from actions in applications
US11074280B2 (en) * 2017-05-18 2021-07-27 Aiqudo, Inc Cluster based search and recommendation method to rapidly on-board commands in personal assistants
WO2018213788A1 (en) 2017-05-18 2018-11-22 Aiqudo, Inc. Systems and methods for crowdsourced actions and commands
US10838746B2 (en) 2017-05-18 2020-11-17 Aiqudo, Inc. Identifying parameter values and determining features for boosting rankings of relevant distributable digital assistant operations
US11340925B2 (en) 2017-05-18 2022-05-24 Peloton Interactive Inc. Action recipes for a crowdsourced digital assistant system
US10768954B2 (en) 2018-01-30 2020-09-08 Aiqudo, Inc. Personalized digital assistant device and related methods
US10803865B2 (en) 2018-06-05 2020-10-13 Voicify, LLC Voice application platform
US11437029B2 (en) * 2018-06-05 2022-09-06 Voicify, LLC Voice application platform
US10235999B1 (en) 2018-06-05 2019-03-19 Voicify, LLC Voice application platform
US10636425B2 (en) 2018-06-05 2020-04-28 Voicify, LLC Voice application platform
WO2020227310A1 (en) * 2019-05-06 2020-11-12 Google Llc Generating and updating voice-based software applications using application templates

Citations (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5737393A (en) * 1995-07-31 1998-04-07 Ast Research, Inc. Script-based interactive voice mail and voice response system
US20010032234A1 (en) * 1999-12-16 2001-10-18 Summers David L. Mapping an internet document to be accessed over a telephone system
US6351679B1 (en) * 1996-08-20 2002-02-26 Telefonaktiebolaget Lm Ericsson (Publ) Voice announcement management system
US20030083882A1 (en) * 2001-05-14 2003-05-01 Schemers Iii Roland J. Method and apparatus for incorporating application logic into a voice responsive system
US20040000647A1 (en) * 2002-06-26 2004-01-01 Horsky Thomas N. Electron impact ion source
US6778643B1 (en) * 2000-03-21 2004-08-17 Sbc Technology Resources, Inc. Interface and method of designing an interface
US6782412B2 (en) * 1999-08-24 2004-08-24 Verizon Laboratories Inc. Systems and methods for providing unified multimedia communication services
US20050028085A1 (en) * 2001-05-04 2005-02-03 Irwin James S. Dynamic generation of voice application information from a web server
US20050132261A1 (en) * 2003-12-12 2005-06-16 International Business Machines Corporation Run-time simulation environment for voiceXML applications that simulates and automates user interaction
US6912691B1 (en) * 1999-09-03 2005-06-28 Cisco Technology, Inc. Delivering voice portal services using an XML voice-enabled web server
US6928156B2 (en) * 2002-05-31 2005-08-09 Sbc Properties, L.P. Automated operator assistance with menu options
US20050203782A1 (en) * 2004-03-12 2005-09-15 Carline Smith Automated airlines reservations system
US20050246174A1 (en) * 2004-04-28 2005-11-03 Degolia Richard C Method and system for presenting dynamic commercial content to clients interacting with a voice extensible markup language system
US7054939B2 (en) * 2001-06-28 2006-05-30 Bellsouth Intellectual Property Corportion Simultaneous visual and telephonic access to interactive information delivery
US7062544B1 (en) * 1999-09-27 2006-06-13 General Instrument Corporation Provisioning of locally-generated prompts from a central source
US7143042B1 (en) * 1999-10-04 2006-11-28 Nuance Communications Tool for graphically defining dialog flows and for establishing operational links between speech applications and hypermedia content in an interactive voice response environment
US7143037B1 (en) * 2002-06-12 2006-11-28 Cisco Technology, Inc. Spelling words using an arbitrary phonetic alphabet
US7206391B2 (en) * 2003-12-23 2007-04-17 Apptera Inc. Method for creating and deploying system changes in a voice application system
US7287248B1 (en) * 2002-10-31 2007-10-23 Tellme Networks, Inc. Method and system for the generation of a voice extensible markup language application for a voice interface process
US7440898B1 (en) * 1999-09-13 2008-10-21 Microstrategy, Incorporated System and method for the creation and automatic deployment of personalized, dynamic and interactive voice services, with system and method that enable on-the-fly content and speech generation
US7817784B2 (en) * 2003-12-23 2010-10-19 Apptera, Inc. System for managing voice files of a voice prompt server

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0769866A3 (en) * 1995-10-19 2001-02-07 Ncr International Inc. Automated voice mail/answering machine greeting system
US7039168B1 (en) * 2002-03-12 2006-05-02 Bellsouth Intellectual Property Corp. System and method for inventory of and access to recorded announcements
US20040247093A1 (en) * 2002-05-20 2004-12-09 Potts Karl W. System, method, and apparatus for provisioning recorded announcements
US7412655B2 (en) * 2004-11-02 2008-08-12 Yahoo! Inc. Method and system of providing dynamic dialogs

Patent Citations (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5737393A (en) * 1995-07-31 1998-04-07 Ast Research, Inc. Script-based interactive voice mail and voice response system
US6351679B1 (en) * 1996-08-20 2002-02-26 Telefonaktiebolaget Lm Ericsson (Publ) Voice announcement management system
US6782412B2 (en) * 1999-08-24 2004-08-24 Verizon Laboratories Inc. Systems and methods for providing unified multimedia communication services
US6912691B1 (en) * 1999-09-03 2005-06-28 Cisco Technology, Inc. Delivering voice portal services using an XML voice-enabled web server
US7440898B1 (en) * 1999-09-13 2008-10-21 Microstrategy, Incorporated System and method for the creation and automatic deployment of personalized, dynamic and interactive voice services, with system and method that enable on-the-fly content and speech generation
US7062544B1 (en) * 1999-09-27 2006-06-13 General Instrument Corporation Provisioning of locally-generated prompts from a central source
US7143042B1 (en) * 1999-10-04 2006-11-28 Nuance Communications Tool for graphically defining dialog flows and for establishing operational links between speech applications and hypermedia content in an interactive voice response environment
US20010032234A1 (en) * 1999-12-16 2001-10-18 Summers David L. Mapping an internet document to be accessed over a telephone system
US6778643B1 (en) * 2000-03-21 2004-08-17 Sbc Technology Resources, Inc. Interface and method of designing an interface
US20050028085A1 (en) * 2001-05-04 2005-02-03 Irwin James S. Dynamic generation of voice application information from a web server
US20030083882A1 (en) * 2001-05-14 2003-05-01 Schemers Iii Roland J. Method and apparatus for incorporating application logic into a voice responsive system
US7054939B2 (en) * 2001-06-28 2006-05-30 Bellsouth Intellectual Property Corportion Simultaneous visual and telephonic access to interactive information delivery
US6928156B2 (en) * 2002-05-31 2005-08-09 Sbc Properties, L.P. Automated operator assistance with menu options
US7143037B1 (en) * 2002-06-12 2006-11-28 Cisco Technology, Inc. Spelling words using an arbitrary phonetic alphabet
US20040000647A1 (en) * 2002-06-26 2004-01-01 Horsky Thomas N. Electron impact ion source
US7287248B1 (en) * 2002-10-31 2007-10-23 Tellme Networks, Inc. Method and system for the generation of a voice extensible markup language application for a voice interface process
US20050132261A1 (en) * 2003-12-12 2005-06-16 International Business Machines Corporation Run-time simulation environment for voiceXML applications that simulates and automates user interaction
US7206391B2 (en) * 2003-12-23 2007-04-17 Apptera Inc. Method for creating and deploying system changes in a voice application system
US7817784B2 (en) * 2003-12-23 2010-10-19 Apptera, Inc. System for managing voice files of a voice prompt server
US20050203782A1 (en) * 2004-03-12 2005-09-15 Carline Smith Automated airlines reservations system
US20050246174A1 (en) * 2004-04-28 2005-11-03 Degolia Richard C Method and system for presenting dynamic commercial content to clients interacting with a voice extensible markup language system
US20110044437A1 (en) * 2004-04-28 2011-02-24 Apptera, Inc. Method and System for Presenting Dynamic Commercial Content to Clients Interacting with a Voice Extensible Markup Language system

Cited By (171)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9646614B2 (en) 2000-03-16 2017-05-09 Apple Inc. Fast, language-independent method for user authentication by voice
US8509403B2 (en) 2003-11-17 2013-08-13 Htc Corporation System for advertisement selection, placement and delivery
US20110044437A1 (en) * 2004-04-28 2011-02-24 Apptera, Inc. Method and System for Presenting Dynamic Commercial Content to Clients Interacting with a Voice Extensible Markup Language system
US7660719B1 (en) * 2004-08-19 2010-02-09 Bevocal Llc Configurable information collection system, method and computer program product utilizing speech recognition
US10318871B2 (en) 2005-09-08 2019-06-11 Apple Inc. Method and apparatus for building an intelligent automated assistant
US8024196B1 (en) * 2005-09-19 2011-09-20 Sap Ag Techniques for creating and translating voice applications
US9117447B2 (en) 2006-09-08 2015-08-25 Apple Inc. Using event alert text as input to an automated assistant
US8942986B2 (en) 2006-09-08 2015-01-27 Apple Inc. Determining user intent based on ontologies of domains
US8930191B2 (en) 2006-09-08 2015-01-06 Apple Inc. Paraphrasing of user requests and results by automated digital assistant
US10568032B2 (en) 2007-04-03 2020-02-18 Apple Inc. Method and system for operating a multi-function portable electronic device using voice-activation
US9330720B2 (en) 2008-01-03 2016-05-03 Apple Inc. Methods and apparatus for altering audio output signals
US10381016B2 (en) 2008-01-03 2019-08-13 Apple Inc. Methods and apparatus for altering audio output signals
US9626955B2 (en) 2008-04-05 2017-04-18 Apple Inc. Intelligent text-to-speech conversion
US9865248B2 (en) 2008-04-05 2018-01-09 Apple Inc. Intelligent text-to-speech conversion
US9535906B2 (en) 2008-07-31 2017-01-03 Apple Inc. Mobile device having human language translation capability with positional feedback
US10108612B2 (en) 2008-07-31 2018-10-23 Apple Inc. Mobile device having human language translation capability with positional feedback
US9959870B2 (en) 2008-12-11 2018-05-01 Apple Inc. Speech recognition involving a mobile device
US8713146B2 (en) * 2009-03-27 2014-04-29 Ebay Inc. Change management automation tool
US20100251247A1 (en) * 2009-03-27 2010-09-30 Ebay Inc. Change management automation tool
US8346560B2 (en) * 2009-05-01 2013-01-01 Alpine Electronics, Inc Dialog design apparatus and method
US20100280819A1 (en) * 2009-05-01 2010-11-04 Alpine Electronics, Inc. Dialog Design Apparatus and Method
US20120081371A1 (en) * 2009-05-01 2012-04-05 Inci Ozkaragoz Dialog design tool and method
US8798999B2 (en) * 2009-05-01 2014-08-05 Alpine Electronics, Inc. Dialog design tool and method
US9858925B2 (en) 2009-06-05 2018-01-02 Apple Inc. Using context information to facilitate processing of commands in a virtual assistant
US11080012B2 (en) 2009-06-05 2021-08-03 Apple Inc. Interface for a virtual digital assistant
US10475446B2 (en) 2009-06-05 2019-11-12 Apple Inc. Using context information to facilitate processing of commands in a virtual assistant
US10795541B2 (en) 2009-06-05 2020-10-06 Apple Inc. Intelligent organization of tasks items
US10283110B2 (en) 2009-07-02 2019-05-07 Apple Inc. Methods and apparatuses for automatic speech recognition
US8670985B2 (en) 2010-01-13 2014-03-11 Apple Inc. Devices and methods for identifying a prompt corresponding to a voice input in a sequence of prompts
US8311838B2 (en) * 2010-01-13 2012-11-13 Apple Inc. Devices and methods for identifying a prompt corresponding to a voice input in a sequence of prompts
US20110172994A1 (en) * 2010-01-13 2011-07-14 Apple Inc. Processing of voice inputs
US10276170B2 (en) 2010-01-18 2019-04-30 Apple Inc. Intelligent automated assistant
US9548050B2 (en) 2010-01-18 2017-01-17 Apple Inc. Intelligent automated assistant
US10705794B2 (en) 2010-01-18 2020-07-07 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US10706841B2 (en) 2010-01-18 2020-07-07 Apple Inc. Task flow identification based on user intent
US8903716B2 (en) 2010-01-18 2014-12-02 Apple Inc. Personalized vocabulary for digital assistant
US8892446B2 (en) 2010-01-18 2014-11-18 Apple Inc. Service orchestration for intelligent automated assistant
US11423886B2 (en) 2010-01-18 2022-08-23 Apple Inc. Task flow identification based on user intent
US9318108B2 (en) 2010-01-18 2016-04-19 Apple Inc. Intelligent automated assistant
US10496753B2 (en) 2010-01-18 2019-12-03 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US10679605B2 (en) 2010-01-18 2020-06-09 Apple Inc. Hands-free list-reading by intelligent automated assistant
US10553209B2 (en) 2010-01-18 2020-02-04 Apple Inc. Systems and methods for hands-free notification summaries
US9633660B2 (en) 2010-02-25 2017-04-25 Apple Inc. User profiling for voice input processing
US10049675B2 (en) 2010-02-25 2018-08-14 Apple Inc. User profiling for voice input processing
US10762293B2 (en) 2010-12-22 2020-09-01 Apple Inc. Using parts-of-speech tagging and named entity recognition for spelling correction
US10102359B2 (en) 2011-03-21 2018-10-16 Apple Inc. Device access using voice authentication
US9262612B2 (en) 2011-03-21 2016-02-16 Apple Inc. Device access using voice authentication
US10706373B2 (en) 2011-06-03 2020-07-07 Apple Inc. Performing actions associated with task items that represent tasks to perform
US10241644B2 (en) 2011-06-03 2019-03-26 Apple Inc. Actionable reminder entries
US11120372B2 (en) 2011-06-03 2021-09-14 Apple Inc. Performing actions associated with task items that represent tasks to perform
US10057736B2 (en) 2011-06-03 2018-08-21 Apple Inc. Active transport based notifications
US9798393B2 (en) 2011-08-29 2017-10-24 Apple Inc. Text correction processing
US10241752B2 (en) 2011-09-30 2019-03-26 Apple Inc. Interface for a virtual digital assistant
US10134385B2 (en) 2012-03-02 2018-11-20 Apple Inc. Systems and methods for name pronunciation
US9483461B2 (en) 2012-03-06 2016-11-01 Apple Inc. Handling speech synthesis of content for multiple languages
US9953088B2 (en) 2012-05-14 2018-04-24 Apple Inc. Crowd sourcing information to fulfill user requests
US10079014B2 (en) 2012-06-08 2018-09-18 Apple Inc. Name recognition system
US9495129B2 (en) 2012-06-29 2016-11-15 Apple Inc. Device, method, and user interface for voice-activated navigation and browsing of a document
US9576574B2 (en) 2012-09-10 2017-02-21 Apple Inc. Context-sensitive handling of interruptions by intelligent digital assistant
US9971774B2 (en) 2012-09-19 2018-05-15 Apple Inc. Voice-based media searching
US10978090B2 (en) 2013-02-07 2021-04-13 Apple Inc. Voice trigger for a digital assistant
US10199051B2 (en) 2013-02-07 2019-02-05 Apple Inc. Voice trigger for a digital assistant
US9368114B2 (en) 2013-03-14 2016-06-14 Apple Inc. Context-sensitive handling of interruptions
US9922642B2 (en) 2013-03-15 2018-03-20 Apple Inc. Training an at least partial voice command system
US9697822B1 (en) 2013-03-15 2017-07-04 Apple Inc. System and method for updating an adaptive speech recognition model
US9582608B2 (en) 2013-06-07 2017-02-28 Apple Inc. Unified ranking with entropy-weighted information for phrase-based semantic auto-completion
US9620104B2 (en) 2013-06-07 2017-04-11 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
US9633674B2 (en) 2013-06-07 2017-04-25 Apple Inc. System and method for detecting errors in interactions with a voice-based digital assistant
US9966060B2 (en) 2013-06-07 2018-05-08 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
US9966068B2 (en) 2013-06-08 2018-05-08 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
US10657961B2 (en) 2013-06-08 2020-05-19 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
US10176167B2 (en) 2013-06-09 2019-01-08 Apple Inc. System and method for inferring user intent from speech inputs
US10185542B2 (en) 2013-06-09 2019-01-22 Apple Inc. Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant
US9300784B2 (en) 2013-06-13 2016-03-29 Apple Inc. System and method for emergency calls initiated by voice command
US10791216B2 (en) 2013-08-06 2020-09-29 Apple Inc. Auto-activating smart responses based on activities from remote devices
US9620105B2 (en) 2014-05-15 2017-04-11 Apple Inc. Analyzing audio input for efficient speech and music recognition
US10592095B2 (en) 2014-05-23 2020-03-17 Apple Inc. Instantaneous speaking of content on touch devices
US9502031B2 (en) 2014-05-27 2016-11-22 Apple Inc. Method for supporting dynamic grammars in WFST-based ASR
US9842101B2 (en) 2014-05-30 2017-12-12 Apple Inc. Predictive conversion of language input
US10169329B2 (en) 2014-05-30 2019-01-01 Apple Inc. Exemplar-based natural language processing
US10078631B2 (en) 2014-05-30 2018-09-18 Apple Inc. Entropy-guided text prediction using combined word and character n-gram language models
US10083690B2 (en) 2014-05-30 2018-09-25 Apple Inc. Better resolution when referencing to concepts
US9633004B2 (en) 2014-05-30 2017-04-25 Apple Inc. Better resolution when referencing to concepts
US10497365B2 (en) 2014-05-30 2019-12-03 Apple Inc. Multi-command single utterance input method
US9430463B2 (en) 2014-05-30 2016-08-30 Apple Inc. Exemplar-based natural language processing
US9966065B2 (en) 2014-05-30 2018-05-08 Apple Inc. Multi-command single utterance input method
US10289433B2 (en) 2014-05-30 2019-05-14 Apple Inc. Domain specific language for encoding assistant dialog
US11257504B2 (en) 2014-05-30 2022-02-22 Apple Inc. Intelligent assistant for home automation
US11133008B2 (en) 2014-05-30 2021-09-28 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US9715875B2 (en) 2014-05-30 2017-07-25 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US10170123B2 (en) 2014-05-30 2019-01-01 Apple Inc. Intelligent assistant for home automation
US9734193B2 (en) 2014-05-30 2017-08-15 Apple Inc. Determining domain salience ranking from ambiguous words in natural speech
US9785630B2 (en) 2014-05-30 2017-10-10 Apple Inc. Text prediction using combined word N-gram and unigram language models
US9760559B2 (en) 2014-05-30 2017-09-12 Apple Inc. Predictive text input
US10904611B2 (en) 2014-06-30 2021-01-26 Apple Inc. Intelligent automated assistant for TV user interactions
US9668024B2 (en) 2014-06-30 2017-05-30 Apple Inc. Intelligent automated assistant for TV user interactions
US10659851B2 (en) 2014-06-30 2020-05-19 Apple Inc. Real-time digital assistant knowledge updates
US9338493B2 (en) 2014-06-30 2016-05-10 Apple Inc. Intelligent automated assistant for TV user interactions
US10446141B2 (en) 2014-08-28 2019-10-15 Apple Inc. Automatic speech recognition based on user feedback
US9818400B2 (en) 2014-09-11 2017-11-14 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US10431204B2 (en) 2014-09-11 2019-10-01 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US10789041B2 (en) 2014-09-12 2020-09-29 Apple Inc. Dynamic thresholds for always listening speech trigger
US9606986B2 (en) 2014-09-29 2017-03-28 Apple Inc. Integrated word N-gram and class M-gram language models
US9886432B2 (en) 2014-09-30 2018-02-06 Apple Inc. Parsimonious handling of word inflection via categorical stem + suffix N-gram language models
US9646609B2 (en) 2014-09-30 2017-05-09 Apple Inc. Caching apparatus for serving phonetic pronunciations
US9668121B2 (en) 2014-09-30 2017-05-30 Apple Inc. Social reminders
US10127911B2 (en) 2014-09-30 2018-11-13 Apple Inc. Speaker identification and unsupervised speaker adaptation techniques
US10074360B2 (en) 2014-09-30 2018-09-11 Apple Inc. Providing an indication of the suitability of speech recognition
US9986419B2 (en) 2014-09-30 2018-05-29 Apple Inc. Social reminders
US10552013B2 (en) 2014-12-02 2020-02-04 Apple Inc. Data detection
US11556230B2 (en) 2014-12-02 2023-01-17 Apple Inc. Data detection
US9711141B2 (en) 2014-12-09 2017-07-18 Apple Inc. Disambiguating heteronyms in speech synthesis
US9865280B2 (en) 2015-03-06 2018-01-09 Apple Inc. Structured dictation using intelligent automated assistants
US9721566B2 (en) 2015-03-08 2017-08-01 Apple Inc. Competing devices responding to voice triggers
US9886953B2 (en) 2015-03-08 2018-02-06 Apple Inc. Virtual assistant activation
US10311871B2 (en) 2015-03-08 2019-06-04 Apple Inc. Competing devices responding to voice triggers
US10567477B2 (en) 2015-03-08 2020-02-18 Apple Inc. Virtual assistant continuity
US11087759B2 (en) 2015-03-08 2021-08-10 Apple Inc. Virtual assistant activation
US9899019B2 (en) 2015-03-18 2018-02-20 Apple Inc. Systems and methods for structured stem and suffix language models
US10614169B1 (en) * 2015-04-07 2020-04-07 West Corporation Script unique prompts
US9672211B1 (en) * 2015-04-07 2017-06-06 West Corporation Script unique prompts
US9842105B2 (en) 2015-04-16 2017-12-12 Apple Inc. Parsimonious continuous-space phrase representations for natural language processing
US10083688B2 (en) 2015-05-27 2018-09-25 Apple Inc. Device voice control for selecting a displayed affordance
US10127220B2 (en) 2015-06-04 2018-11-13 Apple Inc. Language identification from short strings
US10101822B2 (en) 2015-06-05 2018-10-16 Apple Inc. Language input correction
US10356243B2 (en) 2015-06-05 2019-07-16 Apple Inc. Virtual assistant aided communication with 3rd party service in a communication session
US10186254B2 (en) 2015-06-07 2019-01-22 Apple Inc. Context-based endpoint detection
US10255907B2 (en) 2015-06-07 2019-04-09 Apple Inc. Automatic accent detection using acoustic models
US11025565B2 (en) 2015-06-07 2021-06-01 Apple Inc. Personalized prediction of responses for instant messaging
US10747498B2 (en) 2015-09-08 2020-08-18 Apple Inc. Zero latency digital assistant
US11500672B2 (en) 2015-09-08 2022-11-15 Apple Inc. Distributed personal assistant
US10671428B2 (en) 2015-09-08 2020-06-02 Apple Inc. Distributed personal assistant
US9697820B2 (en) 2015-09-24 2017-07-04 Apple Inc. Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks
US10366158B2 (en) 2015-09-29 2019-07-30 Apple Inc. Efficient word encoding for recurrent neural network language models
US11010550B2 (en) 2015-09-29 2021-05-18 Apple Inc. Unified language modeling framework for word prediction, auto-completion and auto-correction
US11587559B2 (en) 2015-09-30 2023-02-21 Apple Inc. Intelligent device identification
US10691473B2 (en) 2015-11-06 2020-06-23 Apple Inc. Intelligent automated assistant in a messaging environment
US11526368B2 (en) 2015-11-06 2022-12-13 Apple Inc. Intelligent automated assistant in a messaging environment
US10049668B2 (en) 2015-12-02 2018-08-14 Apple Inc. Applying neural network language models to weighted finite state transducers for automatic speech recognition
US10223066B2 (en) 2015-12-23 2019-03-05 Apple Inc. Proactive assistance based on dialog communication between devices
US10446143B2 (en) 2016-03-14 2019-10-15 Apple Inc. Identification of voice inputs providing credentials
US9934775B2 (en) 2016-05-26 2018-04-03 Apple Inc. Unit-selection text-to-speech synthesis based on predicted concatenation parameters
US9972304B2 (en) 2016-06-03 2018-05-15 Apple Inc. Privacy preserving distributed evaluation framework for embedded personalized systems
US10249300B2 (en) 2016-06-06 2019-04-02 Apple Inc. Intelligent list reading
US11069347B2 (en) 2016-06-08 2021-07-20 Apple Inc. Intelligent automated assistant for media exploration
US10049663B2 (en) 2016-06-08 2018-08-14 Apple, Inc. Intelligent automated assistant for media exploration
US10354011B2 (en) 2016-06-09 2019-07-16 Apple Inc. Intelligent automated assistant in a home environment
US10490187B2 (en) 2016-06-10 2019-11-26 Apple Inc. Digital assistant providing automated status report
US10733993B2 (en) 2016-06-10 2020-08-04 Apple Inc. Intelligent digital assistant in a multi-tasking environment
US10067938B2 (en) 2016-06-10 2018-09-04 Apple Inc. Multilingual word prediction
US11037565B2 (en) 2016-06-10 2021-06-15 Apple Inc. Intelligent digital assistant in a multi-tasking environment
US10192552B2 (en) 2016-06-10 2019-01-29 Apple Inc. Digital assistant providing whispered speech
US10509862B2 (en) 2016-06-10 2019-12-17 Apple Inc. Dynamic phrase expansion of language input
US10269345B2 (en) 2016-06-11 2019-04-23 Apple Inc. Intelligent task discovery
US11152002B2 (en) 2016-06-11 2021-10-19 Apple Inc. Application integration with a digital assistant
US10297253B2 (en) 2016-06-11 2019-05-21 Apple Inc. Application integration with a digital assistant
US10089072B2 (en) 2016-06-11 2018-10-02 Apple Inc. Intelligent device arbitration and control
US10521466B2 (en) 2016-06-11 2019-12-31 Apple Inc. Data driven natural language event detection and classification
US10553215B2 (en) 2016-09-23 2020-02-04 Apple Inc. Intelligent automated assistant
US10043516B2 (en) 2016-09-23 2018-08-07 Apple Inc. Intelligent automated assistant
US11281993B2 (en) 2016-12-05 2022-03-22 Apple Inc. Model and ensemble compression for metric learning
US10593346B2 (en) 2016-12-22 2020-03-17 Apple Inc. Rank-reduced token representation for automatic speech recognition
US10332518B2 (en) 2017-05-09 2019-06-25 Apple Inc. User interface for correcting recognition errors
US10755703B2 (en) 2017-05-11 2020-08-25 Apple Inc. Offline personal assistant
US11405466B2 (en) 2017-05-12 2022-08-02 Apple Inc. Synchronization and task delegation of a digital assistant
US10791176B2 (en) 2017-05-12 2020-09-29 Apple Inc. Synchronization and task delegation of a digital assistant
US10789945B2 (en) 2017-05-12 2020-09-29 Apple Inc. Low-latency intelligent automated assistant
US10410637B2 (en) 2017-05-12 2019-09-10 Apple Inc. User-specific acoustic models
US10810274B2 (en) 2017-05-15 2020-10-20 Apple Inc. Optimizing dialogue policy decisions for digital assistants using implicit feedback
US10482874B2 (en) 2017-05-15 2019-11-19 Apple Inc. Hierarchical belief states for digital assistants
US11217255B2 (en) 2017-05-16 2022-01-04 Apple Inc. Far-field extension for digital assistant services

Also Published As

Publication number Publication date
US7206391B2 (en) 2007-04-17
US20050135338A1 (en) 2005-06-23
WO2005104472A2 (en) 2005-11-03
WO2005104472A3 (en) 2007-11-08

Similar Documents

Publication Publication Date Title
US7206391B2 (en) Method for creating and deploying system changes in a voice application system
US7817784B2 (en) System for managing voice files of a voice prompt server
US20110044437A1 (en) Method and System for Presenting Dynamic Commercial Content to Clients Interacting with a Voice Extensible Markup Language system
US10534586B2 (en) Method, system and apparatus for visual programming of interaction workflows for omni-channel customer contact centers with integrated customer relationship management
US7286985B2 (en) Method and apparatus for preprocessing text-to-speech files in a voice XML application distribution system using industry specific, social and regional expression rules
US6460057B1 (en) Data object management system
EP1549032B1 (en) Prompt language translation for a telecommunications system and method
US7809376B2 (en) Enhanced analogue of interactive voice response structures and functions for mobile phones and similar handheld communications devices
US20050283764A1 (en) Method and apparatus for validating a voice application
US9069450B2 (en) Multi-modal/multi-channel application tool architecture
US7406418B2 (en) Method and apparatus for reducing data traffic in a voice XML application distribution system through cache optimization
US20100318365A1 (en) Method and Apparatus for Configuring Web-based data for Distribution to Users Accessing a Voice Portal System
US20110299672A1 (en) System and methods for dynamic integration of a voice application with one or more Web services
US8234660B2 (en) Method and apparatus for a support platform
CN101138228A (en) Customisation of voicexml application
CN102263863A (en) Process-integrated tree view control for interactive voice response design
US10922058B2 (en) Method, system and apparatus for visual programming of interaction workflows for omni-channel customer contact centers with integrated customer relationship management
CN113170002B (en) System and method for providing contextual assistance for contact center applications
US9350864B1 (en) Method and system for specifying and processing telephony sessions
US7400718B2 (en) Automated business form information aquisition system
US20040042593A1 (en) Web-based telephony services creation, deployment and maintenance method and system
US6642942B1 (en) Method and system for configuring among call processing applications in a call processing system
US7797676B2 (en) Method and system for switching between prototype and real code production in a graphical call flow builder
US10637958B1 (en) Supporting complex arrays in a data-driven user interface
Bembieva et al. Voice dialogs for Asterisk

Legal Events

Date Code Title Description
AS Assignment

Owner name: HTC CORPORATION, TAIWAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:APPTERA, INC.;REEL/FRAME:029884/0873

Effective date: 20130204

AS Assignment

Owner name: APPTERA, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:CHIU, LEO;EHRLICH, STEVEN CLIVE;REEL/FRAME:030121/0864

Effective date: 20040611

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION