AN ANIMATED, INTERACTIVE COMPUTER INTERFACE SYSTEM Reference to Related Application and Priority Claim
This application expressly claims the benefit of the earlier filing date and right of priority from the following patent application: U.S. Provisional Application Serial No. 60/132,249, filed on May 3, 1999 in the name of Craig Huish and entitled "An Animated,
Interactive Computer Interface System." The entirety of that earlier- filed, co-pending provisional patent application is hereby expressly incorporated herein by reference. Field of Invention
The present invention relates to user interface in computer systems and more specifically, relates to animated, interactive computer interfaces.
Background of Invention
Speech recognition software enables users to interact with personal computers. These speech recognition software programs often integrate with operating software programs and other application software programs so that speech recognition takes place in the background, thereby making speech simply another way of giving the computer commands and instructions. Some speech recognition software programs are designed to control simple functions of the computer's operating program, such as opening and closing of programs, switching from one program to another, moving windows around on-screen, and issuing menu commands. Other speech recognition software programs allow the user to create a library of verbal commands, such as "GET MAIL" to launch an Inbox mail program, dial an e-mail server, and check for new messages on-line, or "FAX" to create and address a new fax form.
Even though most speech recognition software programs are easier to use, more accurate and include greater productivity enhancing tools than the programs of the past, it is believed that many computer users still refuse to use them because they are uncomfortable speaking to a machine. It is believed that one way to overcome this discomfort is to give the computer more human or animal-like characteristics by displaying an animated character on the computer monitor that receives and responds to audio commands inputted by the user.
Some computer application programs employ animated characters to interact with the user. For example, animated characters are used in the word processing program sold under the trademark, WORD 97, by Microsoft Corporation. In this program, a "help" feature
known as "Office Assistance," which displays an animated character when the "help" feature is requested, is provided. However, the activity of the animated character is limited, and the animated character can only perform a simple human or animal-like body movement or facial expression. It is believed that one drawback with application programs that use animated characters is that they do not have speech recognition capabilities. In addition, it is believed that these programs are not designed to provide an audible response to the user's comments or actions. The activities of the animated characters are also not integrated with other operating and application programs to perform functions in those programs. Summary of the Invention
The present invention provides an interactive computer interface system that includes a computer, which has a display monitor and storage medium, a sound receiving device, a sound generating device, and at least one machine executable file stored in the storage medium. A means for converting audio input into machine readable code and an animated, interactive software program are also included in the system. The animated, interactive software program produces a visual and audible response to the audio input and/or performs tasks responsive to the audio input by interpreting the machine readable code.
One embodiment of the animated, interactive software program includes a brain component and an animated character component. The brain component receives and interprets the machine readable code through an operating system and generates a response to the audio input. The animated character component, which is coupled to the brain component, receives the response to the audio input, and then produces the visual and/or audible responses to the audio input.
The present invention also provides an interactive user interface system for a computer that has memory to run computer operating system software and a plurality of software applications. The computer also includes a device that interacts with the memory to create a user input file and an interactive software program resident in the computer memory. The interactive software program has a sensory output component for providing visual and audible output and a brain component for interpreting the user input file, delivering commands for the sensory output component responsive to the user input file, and initiating user-defined tasks for the computer.
The present invention also includes a method of operating an interactive computer interface system. The method includes: providing a computer with a display monitor and a storage medium, a sound generating device, and at least one machine executable file stored in the storage medium; inputting audio to a sound receiving device; converting the audio input into machine readable code; delivering the machine readable code to an animated interactive software program; and executing the at least one machine executable file associated with the machine readable code. Brief Description of the Drawings
The accompanying drawings, which are incorporated herein and constitute part of this specification, illustrate the presently preferred embodiment of the invention, and, together with the general description given above and the detailed description given below, serve to explain the features of the invention.
Fig. 1 is an illustration of the computer system of the interactive computer interface system of the present invention. Fig. 2 is a block diagram of the interactive computer interface system of the present invention.
Fig. 3 is a block diagram of the interactive computer interface system of Fig. 2 connected to a network server.
Detailed Description of the Preferred Embodiment(s) As shown in Fig. 1, the physical aspects of the interactive computer interface system include a computer 1 with a display monitor 2 and a storage medium 3, a sound generating device, such as a speaker 4, and a sound receiving device, such as a microphone 5. The microphone 5 may either be built into the storage medium 3 or stand alone to receive audio input from a user. The system is designed for a computer user to use an animated character, or graphic image, 6 as an assistant when operating the computer 1. As such, the system enables the computer user to operate the computer 1 by inputting speech commands that appear to be received and acted upon by an animated character 6 displayed on the monitor 2.
As shown in Fig. 2, the interactive computer interface system also includes an animated, interactive software program 10, at least one machine executable file 11 stored in the storage medium 3, or computer memory, a means 12 for converting, preferably, audio input 13 into machine readable code 14. Audio input 13 received by the sound recording
devices 5 is preferred, but any other input and devices, such as a mouse, keyboard and touch- sensitive screen, may be used. Preferably, the storage medium 3 has a hard drive (not shown) with sufficient memory to run an operating system resident with the animated, interactive software program 10, the means 12 for converting the audio input 13 into machine readable code, and a plurality of the machine executable files 11, 16, 17, 18 and 19, which are, preferably, compatible with the operating system 15 and the means 12. The machine readable code may be ASCII text files or any other user input file compatible with the operating system 15 on the computer 1. The machine executable files 11, 16, 17, 18 and 19 may be application software programs or executable files associated with the application software programs.
In the preferred embodiment, the animated, interactive software program 10 performs tasks 22 or 23, which may be user-defined, responsive to the audio input 13 by interpreting the machine readable code 14 and/or produces a visual response 25 and audible response 26 to the audio input 13. In the preferred embodiment, the animated, interactive program 10 includes two components: a brain component 20 and an animated character component, or sensory output component, 30. The brain component 20 acts as a central hub between the operating system 15, means 12, or speech recognition program, and the plurality of application software programs 11, 17 and 18 and executable files 16 and 19. The hierarchical structure of the brain component 20, which is similar to the hierarchical structure used in operating systems, accepts propriety software programs requiring minimal storage space, known as "plug-ins," 17 and 18. These plug-ins 17 and 18 plug into the brain component 20 to provide added functionality. Using such a hierarchical structure makes the program 10 expandable or scalable for use with future application programs.
When the computer 1 is booted, or turned on, the operating system 15, the speech recognition program, or means 12, and the brain component 20 are loaded into the computer's random access memory. In the preferred embodiment, speech commands, or audio input 13, are delivered to the computer 1 through the sound receiving device 5, and the speech recognition software program, or means 12, converts the speech, or audio input 13, into machine readable code 14, or standard text files, which are then further processed by the brain component 20. Preferably, the audio input 13 operates through the operating system 15.
When a suitable text file 14 is delivered to the brain component 20 and presented to the
application software program 11 and plug-in programs 17 and 18, the application software program 11 or the single plug-in 17 or 18 associated with the text 14 is activated. In turn, the executable file 16 or 19 associated with the activated plug-in 17 or 18, respectively, may be automatically executed. After the application software program 11 is activated or the executable file 16 or 19 is executed, a set of task instructions 22 or 23 may be generated by the brain component 20. The task instructions 22 may be delivered to the operating system 15, or the task instructions 23 may be delivered to other application programs 11. The set of task instructions 22 or 23 may also be delivered without execution of the application software program 11 or executable file 16 or 19 if the text file 14 is not associated with a plug-in 17 or 18, but associated only with the task instructions 22 or 23. The task instructions 22 or 23 may inform the operating system 15, the application software program 11, or any other program or file to carry out specific tasks, such as opening files or executing software applications. For example, if the user's audio input 13 is "What time is it?", then the task instruction 22 or 23 will be an audible response 26 with the current time and a visual response 25 with visual movements corresponding to the audible response 26. In addition, if the user's audio input 13 is "Open WordPerfect," the program will be opened, and the audible response 26, along with the visual response 25, may be "OK" or "I am opening the program," or something similar with corresponding gestures from the animated character 6. Simultaneously, a set of character instructions 24 are generated by the brain component 20 and delivered to the animated character component 30. These character instructions 24 are used to produce the animated character 6, or graphic image, on the monitor 2 and the audible output 26 from the sound generating device 4. In the preferred embodiment, the visual response 25 and audible response 26 operate through the operating system 15. Alternatively, the visual response 25 and audible response 26 operate stand-alone.
In the preferred embodiment, the animated character component 30 includes a plurality of graphic image files 33 and a plurality of sound files 35. The graphic image files 33 are used to present the human-like body and facial movement features of the animated character 6 on the monitor 2. This animated character 6 may be large enough to fill the entire display of the monitor 2. The animated character 6 may be displayed in two dimensional or three- dimensional animation or photo realistic images and preferably, has shoulders, a neck, and
head. The sound files 35 are used to provide life-like audible responses 26 from the animated character component 30. The graphic image files 33 and sound files 35 may be provided from any program designed to operate with a means 12 to convert audio input 13 into machine readable code 14 and that can be integrated with the operating system 15 and other machine executable files 11, 16, 17, 18 and 19 that may be used. In the preferred embodiment, the animated character component 30 includes a runtime version of an animation program known as DIRECTOR 6.0, sold by Macromedia Inc. of San Francisco, California, and a set of selected graphic image files 33 capable of being executed thereunder. Preferably, the graphic image files 33 are BMP files or PIC files that present a relatively large animated character 6 having human-like body and facial movement features on the monitor 2.
The animated character component 30 also coordinates and generates the audible responses 26 made by the animated character 6. In the preferred embodiment, the audible responses 26 are produced by pre-recorded sound files 35. These sound files 35 may be recorded QUICKTIME, sold by Apple Corporation of California, files or midi or wave formatted files. Alternatively, the audible responses 26 are produced by generic text-to- speech (TTS) audio technology. When the set of character instructions 24 are delivered to the animated character component 30, the sound files 35 associated with the animated activity will automatically be played to corresponding animated character facial (mouth) movements. In the preferred embodiment, the brain component 20 also responds to selected speech commands in the audio input 13. If an unrecognized or a recognized, inappropriate speech command is inputted, the brain component 20 may instruct the animated character component 30 to respond visually or audibly that the input command was unrecognized or inappropriate. When the system is activated and waiting for a speech command to be inputted, the animated character component 30 may be active by displaying an animated character 6 with a slight grin or other small facial movements (i.e., blinking, yawning, etc.). When a specified time period elapses, both the brain component 20 and the animated character component 30 may become inactivated. This may be indicated by the eyes of the animated character 6 being closed or the generation of snoring sounds, thereby indicating that the computer 1 is idle or in a standby mode. In the preferred embodiment, the operating system 15 is WINDOWS 95, WINDOWS
98, or WINDOWS NT, all sold by Microsoft, Inc. of Redmond, WA. The speech recognition
program, or means 12, may also have "command and control" capability so that any function on the computer may be carried out with audio input 13, or speech commands. Some software programs that operate in this manner include DRAGON NATURAL SPEAKING 3.0, sold by Dragon Systems, Inc. of Newton, MA, IBM VIAVOICE 98 EXECUTIVE, sold by IBM Corp. of West Palm Beach, FI, and VOICE XPRESS PLUS 1.01, sold by Lernout &
Hauspie Speech Products USA Inc., of Burlington, MA. However, any compatible program may be employed.
The animated character component 30 may be exchanged by the user to display different animated characters 6 with different voices, depending on the user's preferences. Initially, the brain component 20 and the animated character component 30 are loaded into the computer at the same time. Later, when a new animated character 6 is desired, the user simply removes the old animated character component 30 from the computer 1 and downloads a new animated character component 30a, 30b, or 30c for a new animated character 6 from a CD-ROM or floppy disc or from a network server 50 connected to a wide area network 60, as shown in Fig.3. It should be understood that one or both components 20 and 30 may be resident in the computer 1 or network based, where one or both components 20 and 30 are stored on the network server and remotely controlled by a client machine.
While the invention has been disclosed with reference to certain preferred embodiments, numerous modifications, alterations, and changes to the described embodiments are possible without departing from the sphere and scope of the invention, as defined in the appended claims and their equivalents thereof. Accordingly, it is intended that the invention not be limited to the described embodiments, but that it have the full scope defined by the language of the following claims.