[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

US20070100634A1 - Tracking Time Using Portable Recorders and Speech Recognition - Google Patents

Tracking Time Using Portable Recorders and Speech Recognition Download PDF

Info

Publication number
US20070100634A1
US20070100634A1 US11/566,802 US56680206A US2007100634A1 US 20070100634 A1 US20070100634 A1 US 20070100634A1 US 56680206 A US56680206 A US 56680206A US 2007100634 A1 US2007100634 A1 US 2007100634A1
Authority
US
United States
Prior art keywords
time
determining
voice command
speech
voice
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/566,802
Inventor
James Cooper
Donna Byron
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nuance Communications Inc
Original Assignee
International Business Machines Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by International Business Machines Corp filed Critical International Business Machines Corp
Priority to US11/566,802 priority Critical patent/US20070100634A1/en
Publication of US20070100634A1 publication Critical patent/US20070100634A1/en
Priority to US12/180,797 priority patent/US7664638B2/en
Assigned to NUANCE COMMUNICATIONS, INC. reassignment NUANCE COMMUNICATIONS, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: INTERNATIONAL BUSINESS MACHINES CORPORATION
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G04HOROLOGY
    • G04GELECTRONIC TIME-PIECES
    • G04G21/00Input or output devices integrated in time-pieces
    • G04G21/06Input or output devices integrated in time-pieces using voice
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/10Office automation; Time management
    • G06Q10/109Time management, e.g. calendars, reminders, meetings or time accounting
    • GPHYSICS
    • G07CHECKING-DEVICES
    • G07CTIME OR ATTENDANCE REGISTERS; REGISTERING OR INDICATING THE WORKING OF MACHINES; GENERATING RANDOM NUMBERS; VOTING OR LOTTERY APPARATUS; ARRANGEMENTS, SYSTEMS OR APPARATUS FOR CHECKING NOT PROVIDED FOR ELSEWHERE
    • G07C1/00Registering, indicating or recording the time of events or elapsed time, e.g. time-recorders for work people
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems

Definitions

  • the present invention relates to time processing and, more particularly, relates to tracking time using portable recorders and speech recognition.
  • Time is a fundamental constraint for all human endeavors, and the primary source of revenue for businesses that charge clients per unit of time. Knowing where time is spent is crucial for any productivity analysis, whether on a professional or a personal level. Manual methods for capturing time, such as journals, time cards and logs, were used by professionals for many years, but in the recent past automated tools have been developed to support this process.
  • time increment denotes a time segment during which a particular person was dedicated to a particular task.
  • a person then creates time increment records, typically by using one of two methods.
  • time increments are recorded after the fact based on estimates.
  • An example of this method is a project management software package that allows its users to allocate time spent on tasks over a large time interval, perhaps a whole week.
  • consultants working on a software project use the project management software at the end of the workweek to record the tasks to which they dedicated time during the previous week.
  • there is no direct capture of the time spent and the information generated is generally imprecise.
  • time increment data is captured directly by interacting with software running on a desktop or hand-held computer platform that allows the user to start and stop a timer at the same time that the task starts and ends.
  • This “real-time” capture is much more accurate than the previous method, because the time increment record is captured when the task is performed and an automatic clock is used to compute the interval rather than a human estimate.
  • using a computer as a capture device is inconvenient for a variety of reasons that will be discussed below.
  • time increment records can be summarized in order to generate a variety of productivity analyses or billing reports.
  • an inaccurate capture method such as recording hours at the end of the week, any analysis generated is purely an approximation.
  • Interacting with a computer whether a hand-held or desktop machine, generally requires a complete interruption to the task at hand and typically takes several seconds to perhaps minutes to complete.
  • the interruption of workflow and the amount of time required becomes a barrier to using computer-based technologies.
  • Such interfaces tend to be used for minimal data capture, i.e., only capturing billable time or only some of the actual billable time. This limits the usefulness of such products for other sorts of time analysis, for example, personal productivity analysis, time analysis for non-billable workers in support roles, and, most importantly, capturing and analyzing where non-billable time is being spent by billable workers.
  • These ease-of-use barriers also prevent professionals from capturing small increments of time spent on billable projects, such as telephone calls made from the airport or electronic mail messages composed while on a commuter train. The revenue that would otherwise be generated from such activities is therefore lost.
  • Using a portable computer device for capture gets past the problem of portability, but introduces its own set of problems.
  • Many professionals who wish to keep their interaction with computers to a minimum do not feel comfortable using such devices, and others cannot afford them.
  • Even those people with access to portable computer devices face many ease-of-use problems. All such devices require two-handed operation and cannot be operated while walking. The entry of a time increment therefore requires the operator to completely interrupt the normal flow of his or her task.
  • the present invention provides for better and simplified time tracking by using portable recorders and speech recognition. This increases accuracy, as more time should be tracked in real-time, and ease of use, as portable recorders are simple, inexpensive and convenient.
  • the present invention converts speech, preferably recorded on a portable recorder, to text, analyzes the text, and determines voice commands and times when the voice commands occurred.
  • Task names are associated with voice commands and time segments. These time segments and tasks may be packaged as time increments and stored (e.g., in a file or database) for further processing.
  • phrase grammar rules are used when analyzing the text, as this helps to determine voice commands. Using phrase grammar rules also allows the text to contain a variety of topics, only some of which are pertinent to tracking time.
  • FIG. 1 is a system for tracking time using portable recorders and speech recognition, in accordance with one embodiment of the present invention
  • FIG. 2 is a flowchart of a method for tracking time using portable recorders and speech recognition, in accordance with one embodiment of the present invention
  • FIG. 3 is a flowchart of a method for converting speech data into time increment records, in accordance with one embodiment of the present invention
  • FIG. 4 is a flowchart of a method for determining and creating time increments, in accordance with one embodiment of the present invention
  • FIG. 5 illustrates a portion of speech data
  • FIG. 6 provides an illustration of determining time increments from voice commands and times for those commands, in accordance with one embodiment of the present invention
  • FIG. 7 illustrates phrase grammar rules in accordance with one embodiment of the present invention.
  • FIG. 8 shows a user interface for showing, analyzing and correcting time increments, in accordance with one embodiment of the present invention.
  • the present invention allows speech recorded on a portable recorder to be converted to time increments.
  • Each time increment can contain a time segment and a task name.
  • other information such as a matter number or comments may also be packaged into a time increment.
  • the present invention will act to supplement existing time and billing programs. For instance, a person could use the present invention to create time increments from speech and then import the time increments into a time and billing program.
  • the present invention can fill a niche that currently remains unfilled.
  • FIG. 1 shows a block diagram of an exemplary system 100 for tracking time using portable recorders and speech recognition.
  • System 100 comprises a portable speech recorder 110 and a computer system 105 .
  • Portable speech recorder 110 comprises speech data 119 .
  • Computer system 105 is any type of computer system, such as a personal computer, and it comprises processor 120 and memory 125 .
  • Speech data 119 comprises speech files 111 and 115 .
  • Speech file 111 comprises a time stamp 112 , a date stamp 113 , and speech 114 .
  • speech file 115 comprises a time stamp 116 , a date stamp 117 , and speech 118 .
  • Memory 125 comprises speech data 130 , a time-logging method 140 , a speech recognition engine 145 , vocabulary 150 , word hypotheses 151 , text output 155 , time information 160 , phrase grammar rules 165 , voice commands 170 , a time record 175 , a user interface 180 , a time and billing system 185 , and an alternative format time record 190 .
  • Speech data 130 comprises speech files 131 and 135 .
  • Speech file 131 comprises a time stamp 132 , a date stamp 133 , and speech 134 .
  • speech file 135 comprises a time stamp 136 , a date stamp 137 , and speech 138 .
  • Time record 175 comprises time increments 166 and 167 .
  • Computer system 105 can receive computer-readable code means from sources such as compact disk 107 .
  • the methods and apparatus discussed herein may be distributed as an article of manufacture, such as compact disk 107 , that itself comprises a computer readable medium having computer readable code means embodied thereon.
  • the computer readable program code means is operable, in conjunction with a computer system such as computer system 105 , to carry out all or some of the steps to perform the methods or create the apparatuses discussed herein.
  • the computer readable medium may be a recordable medium (e.g., floppy disks, hard drives, compact disks, or memory cards) or may be a transmission medium (e.g., a network comprising fiber-optics, the world-wide web, cables, or a wireless channel using time-division multiple access, code-division multiple access, or other radio-frequency channel). Any medium known or developed that can store information suitable for use with a computer system may be used.
  • the computer-readable code means is any mechanism for allowing a computer to read instructions and data, such as magnetic variations on a magnetic medium or height variations on the surface of a compact disk.
  • Memory 125 will configure the processor 120 to implement the methods, steps, and functions disclosed herein.
  • the memory 125 could be distributed or local and the processor could be distributed or singular.
  • the memory 125 could be implemented as an electrical, magnetic or optical memory, or any combination of these or other types of storage devices.
  • the term “memory” should be construed broadly enough to encompass any information able to be read from or written to an address in the addressable space accessed by processor 120 . With this definition, information on a network is still within memory 125 because the processor 120 can retrieve the information from the network. It should be noted that if processor 120 is distributed, there could be multiple address spaces, one address space per processing unit that makes up processor 120 .
  • Portable recorder 110 is any type of recorder that can store speech.
  • portable recorder 110 records a time stamp for each amount of speech that is recorded.
  • this time stamp indicates when the speech starts, but the time stamp could also indicate when the speech ends.
  • portable recorder 110 contains speech data 119 .
  • Speech data 119 contains a number of speech files, of which speech files 111 and 115 are shown. Speech data 119 could be stored on a tape or other memory device.
  • speech files contain time and date stamps along with speech.
  • speech file 111 contains time stamp 112 , date stamp 113 and speech 114
  • speech file 115 contains time stamp 116 , date stamp 117 and speech 118 .
  • the time and date stamps could be speech, vocalized by an operator of the portable recorder 110 , of the time and date.
  • the time and date stamps are determined by a clock (not shown) of the portable recorder 110 and are placed onto the speech data 119 by portable recorder 110 .
  • Speech 111 and 115 is analog or digital information of the voice of a person.
  • the speech 111 , 115 is digital information that can be easily loaded into computer system 105 .
  • Analog speech may also be loaded into a computer system, but it is usually converted to digital speech.
  • speech 114 and 118 is digital speech, it could be compressed or uncompressed, although, depending on the speech recognition engine used, it may have to be decompressed to be used with a speech recognition engine.
  • the portable recorder 110 is a digital recorder such as the Olympus D1000 digital voice recorder.
  • This recorder digitally stores speech information, and it can add date and time stamps to speech.
  • the speech is stored in a format known as Digital Speech Standard (DSS), which provides for compression of the speech.
  • DSS Digital Speech Standard
  • the speech can be decompressed through a DSS program (not shown in FIG. 1 ) provided as part of a package with the D1000.
  • this functionality could be built into time-logging method 140 , to be discussed below.
  • a digital recorder is preferred for several reasons.
  • some analog media, such as tapes degrade fairly quickly.
  • tapes do not contain very good quality speech recordings. Consequently, speech recognition tends to be poor with tape recordings.
  • digital recorders are better at retaining a high quality of speech. This will aid in speech recognition.
  • Another consideration is the overall quality of speech on a portable recorder. If a portable recorder is used in a quiet environment, such as an office having a door and no background noise, then even the relatively poor quality of speech recorded on a tape may be converted to text with a relatively high degree of accuracy. In general, however, a portable recorder will be used in many places where the background or ambient noise is high. The voice of the speaker in these situations will be relatively hard to determine.
  • a digital portable recorder provides the best recording of speech under such conditions, and digital recorders can contain special circuits and software/firmware to reduce background noise.
  • a final consideration is that a digital portable recorder can place time and date stamps on the speech data. This makes the determination of times easier because the times do not have to be converted to text, which always has a probability of error.
  • Portable recorder 110 in embodiments of the present invention that use digital portable recorders, can be any type of device that can digitally record speech and that can provide time stamps for the speech.
  • a portable recorder 110 could be a digital personal assistant with a program to allow speech to be recorded and time stamped.
  • portable recorder 110 is preferably a digital voice portable recorder.
  • Speech data 130 is a representation of speech data 119 , after speech data 119 has been loaded into computer system 105 . If speech data 119 is analog speech data, the computer system 105 can digitize this data before storing it as speech data 130 . If portable recorder 110 is a digital voice portable recorder, then speech data 130 will be the representation of speech data 119 that is loaded to the computer system 105 by the portable recorder 110 and any program associated with transferring the speech data 119 to speech data 130 . Speech data 130 has a number of speech files 131 through 135 . Speech file 131 has time stamp 132 , date stamp 133 and speech 134 . Preferably, the time stamp 132 and date stamp 133 are digital representations of the time and date, respectively.
  • the time stamp 132 should correspond to time stamp 112
  • date stamp 133 should correspond to date stamp 113
  • speech 134 should correspond to speech 114
  • Speech file 135 has time stamp 136 , date stamp 137 and speech 138 .
  • the time stamp 136 and date stamp 137 are digital representations of the time and date, respectively.
  • the time stamp 136 should correspond to time stamp 116
  • date stamp 137 should correspond to date stamp 117
  • speech 138 should correspond to speech 118 .
  • the speech data 130 could be in a proprietary format. However, the time stamps 132 , 136 and date stamps 133 , 137 can still be accessed, deciphered and correlated with the appropriate speech 134 , 138 . Furthermore, as discussed below, the speech can still be accessed for conversion to text.
  • Time-logging method 140 controls the computer system 105 to convert speech in speech data 130 to time increment record 175 .
  • time-logging method may also provide a user interface 180 (to be discussed below) and can facilitate transfer of speech data 119 into speech data 130 .
  • Time-logging method 140 can contain some, many or all of the steps necessary to perform the methods and create the apparatus or rules discussed below in reference to FIGS. 3 through 8 .
  • the time-logging method 140 accesses or interfaces with the speech recognition engine 145 to cause the engine to convert speech in speech data 130 to text.
  • the speech recognition engine 145 uses a vocabulary 150 to determine appropriate word hypotheses 151 .
  • the speech recognition engine 145 selects the word hypothesis with the highest probability from the word hypotheses 151 to create text output 155 .
  • Text output 155 could contain multiple text files. For instance, there could be a text file corresponding to speech 134 and another text file corresponding to speech 138 . Alternatively, there could be one large text file that contains all the text, perhaps with demarcations to indicate snippets of text and different start times.
  • the speech recognition engine 145 also produces time information 160 that allows the text in text output 155 to be tracked relative to a starting time. As explained in more detail in reference to FIG. 5 , time information 160 is beneficial to determine when a voice command was said relative to a time stamp.
  • Speech recognition engine 145 can be any of a number of speech recognition engines, such as the speech recognition engine that comes with VIAVOICE, which is a popular speech recognition program from IBM (International Business Machines, Incorporated, New Orchard Road, Armonk, N.Y., 10504). As discussed above and discussed in more detail below in reference to FIG. 6 , it is preferred that the speech recognition engine used provide time information that relates the time that text occurs with a relative time of speech. The speech engine VIAVOICE makes this type of timing information available. A program written in the TCL language can be written to access the time information from VIAVOICE, and this information will be relative to the beginning of a speech file.
  • Time-logging method 140 can then compare the words in the text output 155 with phrase grammar rules 165 to determine voice commands 170 .
  • the phrase grammar rules 165 allows system 100 to specifically look for certain key words or terms. These key words and terms are the particular voice commands for which the system is trained.
  • the voice commands are speech snippets that indicate to computer system 105 that the user wants time logging performed. For instance, the user could use the voice commands “going home,” “starting clientA,” or “ending clientA.” These are terms that the computer system 105 should find and decipher to be able to log time.
  • the time-logging method 140 uses the voice commands 170 to create time record 175 , which contains time increments 176 through 177 . Each time increment, as shown in FIG. 6 below, will preferably contain a time segment and a task name. The task name will preferably be a nickname that can be expanded, if desired, by time-logging method 140 . This is discussed in more detail in reference to FIG. 8 .
  • User interface 180 is an optional but preferred component of time-logging method 140 .
  • user interface 180 will be part of a software package (not shown) that contains time-logging method 140 , and user interface 180 may be made integral to time-logging method 140 .
  • User interface 180 allows an operator to perform analyses on the time increments 176 , 177 of time record 175 .
  • the user interface can allow the user to enter or change data in the time record 175 .
  • the user interface 180 may also convert the time record 175 to a form suitable to importing into a time and billing system, such as time and billing system 185 .
  • FIG. 1 an alternative format time record 190 is shown that has been converted from time record 175 into a form suitable for importation into time and billing system 185 .
  • TIMESLIPS is a time and billing program owned by Sage U.S. Holdings, Incorporated, a wholly-owned subsidiary of The Sage Group, Plc., based in Newcastle-upon-Tyne, England.
  • TIMESLIPS contains an importation program called TSIMPORT that can import data from a source file into a TIMESLIPS database.
  • TSIMPORT can import tab-delimited files (where fields are separated by tabs), comma-delimited files (where fields are separated by commas), custom-delimited files (where fields are separated by delimiters specified by the user), and other types of files.
  • the user interface 180 allows a user to convert time record 175 into one or more of these types of files.
  • time record 175 it is possible to create time record 175 directly in one of these types of file formats such that no conversion need take place.
  • the user interface 180 is still beneficial, as it allows correcting and analyzing data in time record 175 prior to importing the time records into time and billing system 185 .
  • FIG. 1 shows a system 100 that allows a person to record voice on a portable recorder and that creates a time record from spoken commands.
  • System 100 is convenient because all a user generally has is a portable recorder into which the user dictates.
  • the conversion from speech data to time increments and time records can occur with any computer system having the appropriate software installed.
  • computer system 105 could be a network ed server and speech data 119 could be loaded onto the computer system 105 through a network.
  • speech recognition engine 145 it is also possible for speech recognition engine 145 , vocabulary 150 , and word hypotheses 151 to be part of portable recorder 110 . This could occur, e.g., if the portable recorder is a personal digital assistant that can support speech processing. In this case, the text output 155 would be uploaded to the computer system 105 for processing. Also, time information 160 could be determined by having the computer system 105 query the speech recognition engine 145 (now on the portable recorder 110 ) for the time information 160 or having the portable recorder 110 upload the time information 160 to the computer system 105 .
  • FIG. 2 shows a flow chart of an overall method 200 for tracking time using portable recorders and speech recognition, in accordance with one embodiment of the present invention.
  • Method 200 is performed whenever a computer system is being prepared to convert speech into time increments. Once the computer system is prepared to convert speech into time increments, then some steps (as discussed below) of method 200 may have already been performed.
  • Method 200 begins when speech recognition and time logging software are installed on a computer system (step 210 ). This step generally involves using an article of manufacture, such as a compact disk, to configure a system to run the speech recognition and time logging software. Step 210 is used to prepare a computer system to recognize speech and change speech into time increments. Once the computer system is prepared, step 210 generally does not need to be performed again.
  • an article of manufacture such as a compact disk
  • a user can create time-tracking categories by using a user interface.
  • time-tracking categories could include a task name (or nickname) and other information pertinent to a task, such as billable rate, comments, and billing category or matter number.
  • the time-tracking categories may be imported from a currently existing time and expense tracking program. Time-tracking categories will be discussed below in reference to FIG. 8 .
  • step 230 the user trains the speech recognition software on the pronunciation of voice commands.
  • the present invention can be used with many speech recognition engines, and these engines could be solely used to convert speech to text for time tracking only. If this is the case, then the speech recognition software can have a limited vocabulary. Training the speech recognition software is always beneficial, regardless of the size of the intended vocabulary, as different people will pronounce the same words differently. Alternately, the speech recognition software could be generally trained for a variety of words, unrelated to the voice commands. It is still beneficial to specifically train the speech recognition software for voice commands, as it is unlikely that the vocabulary of the speech recognition software would contain some of the relatively specific information in a voice command, such as a task name (e.g., “clientA”).
  • step 230 may not have to be performed again. This is particularly true if the speech recognition system can learn new words through feedback. For instance, step 230 could be periodically run to add new voice commands or tracking categories to the system.
  • step 240 the user creates speech data on a portable recorder by speaking voice commands into the portable recorder.
  • the user may also speak additional information, such as dictated letters, grocery lists or important thoughts, onto the portable recorder.
  • the present invention can separate the voice commands from the other speech.
  • step 250 the user loads the speech data onto the computer system. This could be performed through any mechanism known to those skilled in the art, such as a wireless infrared system, serial cables, a universal serial bus connection, or a firewire connection.
  • the speech data is converted into time increment records, which can be packaged into a time record. This is discussed in more detail in FIG. 3 , below.
  • step 270 the user has the option of loading the time increment records into a user interface for review, correction or analysis.
  • An exemplary user interface is shown in FIG. 8 .
  • step 280 the user has the additional option of converting, if necessary, the time records into a file suitable for importation into a time and billing software package. This allows the time increment records created with the present invention to be used with existing software.
  • FIG. 3 shows a flow chart of an exemplary method 260 for converting speech data into time increment records.
  • Method 260 is performed whenever a user would like speech converted to time increment records. Generally, this occurs after the user speaks voice commands into a portable recorder for period of a day or several days. The user then loads the speech data containing the voice commands into a computer system, which then performs method 260 .
  • Method 260 begins when the acoustic signal, which represents speech 320 , is compared to predefined vocabulary words from a vocabulary such as vocabulary 150 .
  • a vocabulary is common in speech processing.
  • a speech recognition engine can create recognized word hypotheses (step 330 ), which are essentially hypotheses, having certain probabilities, that particular words have been said. From the recognized word hypotheses, the speech recognition engine selects the most probable words. Thus, at the end of step 330 , the speech recognition engine has changed speech into text.
  • phrases are composed from words.
  • Phrase grammar rules 165 are used during this step to help in determining phrases associated with voice commands.
  • the voice commands are extracted from the phrases and indexed. Extracting the voice commands could be simple, because the phrases should be highly representative of voice commands. However, there are times, as explained in more detail below in reference to FIG. 5 , when additional processing might be needed to distinguish terms used in voice commands from the same terms used in normal speech. In other words, there could be times when the context in which a term is used will determine whether the term is or is not a voice command. It is possible to use a portable recorder only for tracking time and only for recording voice commands. In this situation, context should not be important.
  • the voice commands are indexed by determining the time, and also preferably the date, when the voice commands were recorded. As explained in more detail in reference to FIG. 6 below, this allows time increments to be created with time segments.
  • the time increments are created. If the speech is long and has many voice commands on it, the number of time increments could also be quite long.
  • the time increments are created, they are written to long- or short-term memory for further processing. They may also be packaged into a time record, which can be a file having a specific, binary, or ASCII format.
  • Method 360 begins when a voice command is selected. Generally, this would be a voice command at the beginning of a list of such commands.
  • step 410 it is determined if the voice command contains a start marker, which is a marker before or after a task name, that indicates that a task is starting.
  • start markers can occur before or after a task name or could, optionally but not preferred, be task names. For example, in the voice commands “start clientA” and “clientA starting,” the words “start” and “starting” are start markers and “clientA” is a task name.
  • active it is meant that, according to the immediately preceding voice command, a task associated with the preceding voice command has been started. For instance, if the previous voice command was the statement, converted to text, “begin ClientA,” then the task associated with the task name ClientA is currently being performed.
  • the active task YES
  • a place holder name such as “unknown” could be used. For example, if the current voice command is “start” but there is no task name, then “unknown” might be used. This could occur, for instance, in cases where the task name is converted to different text or is unintelligible. If the next voice command is “end ClientA,” then the place holder name could be changed to ClientA.
  • steps 420 , 425 and 430 allow starting voice commands to end the previously active task.
  • the task “clientA” will be an active task when the subsequent voice command is issued.
  • the subsequent voice command actually starts a different task.
  • the present invention will stop the task of “clientA” and start the subsequent task (“clientB,” lunch, meeting, or home/end of day).
  • an ending voice command is not required to end the currently active task. This is in marked contrast to current time tracking systems, which do not and cannot assume that one task is ending if another task has begun.
  • the present invention can “intelligently” determine when a currently active task should end.
  • step 468 YES
  • step 480 the active task is ended (step 480 ) and a time increment created for the active task (step 485 ).
  • the portable recorder is a digital voice recorder that contains time stamps along with speech.
  • the content of the speech for this example will be “start clientA . . . end . . . start clientB start lunch . . . stop clientC,” where the ellipses indicate additional spoken words, pauses or stops
  • the present invention using grammar rules such as those discussed below in reference to FIG. 7 , will determine that the voice commands are “start clientA,” “end,” “start clientB,” “start lunch,” and “stop clientC.”
  • the method starts again on step 405 .
  • the selected voice command is “start clientB.”
  • the time from when the previous voice command was recorded to the time of the current voice command is marked as unclassified (step 440 ) and a time increment is created (step 440 ) for the unclassified time.
  • Steps 450 , 490 and 405 are performed.
  • the next voice command is selected (step 405 ) and is the “start lunch” voice command.
  • There is an active task (the task of clientB) in step 415 , so the active task is ended (step 420 ) and a time increment created (step 425 ) for clientB
  • step 465 YES
  • step 468 NO
  • step 470 A past time increment is created (step 470 ) for the task of “lunch,” and a current time increment is created for the task of “clientC” (step 475 ).
  • step 495 There are no more voice commands, so the method ends in step 495 .
  • Method 360 thus attempts to provide time increments for the maximum number of voice commands. Additional steps may be added or current steps modified, by those skilled in the art, to ensure that unusual combinations of voice commands will be captured correctly by method 360 .
  • Speech data 500 comes from a digital voice recorder that records time and date stamps.
  • the digital voice recorder in this example is being used to track time and as a dictation recorder.
  • Speech data 500 comprises time stamps 505 , 530 , 535 , 555 , and 580 , date stamps 510 , 531 , 540 , 560 , 585 , and speech 515 , 532 , 545 , 565 , and 590 .
  • the digital voice recorder records speech data 500 , it places time and date stamps whenever the recorder is started (which can include returning from a pause). For example, the person using this recorder pressed the “record” button and spoke the words “start clientA review and change letter to client.” The recorder created the time 505 and date 510 stamps.
  • Speech 515 contains a voice command 520 and notes 525 concerning the voice command 520 .
  • Speech 532 simply contains a voice command.
  • Speech 545 contains a voice command 550 and a note about the voice command 550 .
  • Speech 565 contains the beginning of a dictated patentability opinion letter, while speech 590 contains the end of the dictated letter and a voice command 5
  • the present invention When the present invention examines this speech data 500 , the present invention will determine that speech 520 , 550 and 595 are voice commands but that speech 570 is not. The present invention can determine this because the phrase grammar rules, as shown and discussed below with reference to FIG. 7 , do not allow a task name by itself to start or stop a task. In speech 570 , there are no other start or end markers around the word “clientB,” and the present invention, based on phrase grammar rules, will determine that there is no voice command in speech 590 .
  • Voice command 520 is associated with time stamp 505 , or 9:15. In other words, the voice command 520 is assumed to have been said at about this time. Similarly, voice commands 532 and 550 are also assumed to have been said at about the times of time stamps 530 (9:25) and 535 (9:30), respectively. However, voice command 595 is said quite a bit of time after time stamp 580 . To determine the time that voice command 595 was said, the present invention preferably accesses time information given by the speech recognition system. The time information should relate the text with a relative time after the speech has started. In general, speech 525 , 532 , 545 , 565 and 590 can be considered separate files A speech recognition engine should treat each separately and should separately track time for each.
  • the time information should start for each at the beginning of the speech. For instance, relative time 593 , determined by accessing the time provided by the speech recognition engine, indicates that 4 minutes and 55 seconds have elapsed since the beginning of speech 590 . This relative time 593 can be added to the time in time stamp 580 to determine that the voice command 595 was said at about 12:15 Thus, the relative time information from a speech recognition engine can be helpful to determine when a voice command was issued.
  • this type of relative time information should not be needed if a digital portable recorder is used solely to track time, if the person stops and starts a digital recorder immediately prior to issuing a voice command, or if an analog portable recorder is used and the person states the time along with a voice command.
  • FIG. 6 shows a table 600 that contains voice command information 610 and time increment information 650 .
  • Voice command information 610 is compiled after the speech has been converted to text and phrase grammar rules have been used to determine voice commands.
  • the content of the speech, which has been changed into voice command information, in FIG. 6 is most of the speech shown in FIG. 5 .
  • the time increment information can be determined. For instance, method 360 of FIG. 4 could be used to determine the time increment information 650 .
  • voice command information 610 comprises voice command entries 615 , 620 , 625 , 630 , and 635 .
  • Time increment information 650 comprises time increments 655 , 660 , 665 , 670 , and 675 .
  • Each voice command entry comprises a time 680 associated with a voice command 685 .
  • the present invention creates time increment information 650 from the voice command information 610 .
  • Many of the entries in the example in FIG. 6 have also been discussed previously in reference to FIG. 4 .
  • Time increment 655 is created from voice command entries 615 and 620 .
  • Time increment 660 is created from voice command entries 625 and 620 .
  • Time increment 665 is created from voice command entries 630 and 625 . It is assumed that starting a current task prior to ending a previous task will stop the previous task and start the current task. If desired, it is possible that each task must be stopped prior to starting another task. In this situation, the voice command entry 630 could cause an error time inclement.
  • Time increments 670 and 675 are created from voice command entries 630 and 635 . There is an error because there is no indication as to when lunch stopped and when the work for clientC began.
  • silence may be an indication that the current task is to end Fox instance
  • quitting time is 6 pm
  • the system could record the ending time as 6 pm even though no ending voice command was given
  • the system could indicate 6 pm as the ending time but indicate a possible error in the time increment. Such an error could be indicated as “6:00?” or “6:00!”
  • phrase grammar rules 165 that are used to determine voice commands from a text file.
  • the rules shown in FIG. 7 are only examples of possible phrase grammar rules, and the actual set of rules used in an embodiment will probably include more rules than those shown, could include additional lexical terms, and could be differently written. There are a variety of ways of writing rules suitable for use with the present invention, and the rules included herein are merely examples of one way to do this.
  • Phrase grammar rules 165 contain indications of tasks and their associated task names in the ⁇ task> category. The symbol “
  • the rules 165 contain start markers, which indicate that a task is beginning.
  • the start markers can be said before a task name, which is the ⁇ start-premarker>, or said after a task name, which is the ⁇ start-postmarker>.
  • the rules 165 also contain end markers, which indicate that a task is ending.
  • the end markers can be said before a task name, which is the ⁇ end-premarker>, said after a task name, which is the ⁇ end-postmarker>, or said alone, which is the ⁇ end-maker>.
  • matter numbers which are task-specific numbers that track more specific tasks. For instance, a matter number of 900 might indicate patentability opinions, which are tasks more specific than just general patent tasks. In the example of FIG.
  • client-specific tasks may also be tracked. These tasks may be tracked by number or words
  • a client-specific task might be, e.g., “clientA 9765,” where “9765” indicates a different task than, say, “9764.”
  • the task “9765” might correspond to “trademark application,” while “9764” might correspond to “patent litigation.”
  • phrases are series of words that indicate that a task has started or ended. For instance, one of the allowed phrases is indicated in ⁇ phrase> as “ ⁇ start-premarker> or ⁇ end-premarker> and ⁇ task>.” Using this phrase definition, approved phrases are “start projectA,” “end lunch” or “quitting meeting.” Then the system can determine whether the phrase is starting or stopping the current task by determining if a start or end, respectively, marker is used.
  • phrase grammar rules 165 are thus specific to the types of words used to start and end tasks.
  • the list of phrase grammar rules 165 may be expanded at any time, and any new words added to the rules 165 may be trained in a speech recognition engine to provide for higher possibility that the new words will be correctly converted to speech.
  • phrase grammar rules 165 are, on the other hand, rules that define what words can be put together (and in what order) to create valid voice commands, but the phrase grammar rules are not the voice commands themselves.
  • an initial set of predefined phrase grammar rules will be included with the system, and interface 180 will allow the user to modify, delete, or add new rules.
  • FIG. 8 this figure shows an exemplary user interface 800 that allows a person to view, analyze, modify, export, create, and remove time increments.
  • the user interface 800 can be created through a number of graphical user interfaces, as is known in the art.
  • User interface 800 has a file menu (File, Consultant, Project, Time Records, Billing), a time record, billable activities, and non-billable activities.
  • the time record shows the current time increments that are being analyzed or viewed.
  • the “nickname” is a task name that indicates a task.
  • the billable activities section shows a nickname, description and billing late.
  • the non-billable activities section shows a nickname and a description.
  • the nickname, description and billing late can be used to analyze the time inclement data and/or could be output to a file for subsequent loading into time and billing software such as TIMESLIPS.
  • the description can be further improved to add additional information.
  • a nickname of “kraft900” could have a description of Kraft, matter 900, patentability opinion.
  • This additional data could be output to a file for further processing by time and billing software.
  • a user can freely add and change the interface, and can use the interface to output files suitable for loading into a time and billing program. For example, under “time records,” an option might be “export,” which allows the further selection of different types of exportable file formats and which data, if any, from billable and non-billable activities sections will supplement or replace the time increments.
  • time tacking system that can convert speech on portable recorders to time increments that can then be analyzed, changed or viewed.
  • the time tracking system can create an output file having time increments that allow a user to import the file into a separate time and expense billing program.

Landscapes

  • Engineering & Computer Science (AREA)
  • Business, Economics & Management (AREA)
  • Physics & Mathematics (AREA)
  • Human Resources & Organizations (AREA)
  • General Physics & Mathematics (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Strategic Management (AREA)
  • Tourism & Hospitality (AREA)
  • Theoretical Computer Science (AREA)
  • Quality & Reliability (AREA)
  • Marketing (AREA)
  • Economics (AREA)
  • General Business, Economics & Management (AREA)
  • Data Mining & Analysis (AREA)
  • Operations Research (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Time Recorders, Dirve Recorders, Access Control (AREA)
  • Telephonic Communication Services (AREA)

Abstract

In general, the present invention converts speech, preferably recorded on a portable recorder, to text, analyzes the text, and determines voice commands and times when the voice commands occurred. Task names are associated with voice commands and time segments. These time segments and tasks may be packaged as time increments and stored (e.g., in a file or database) for further processing. Preferably, phrase grammar rules are used when analyzing the text, as this helps to determine voice commands. Using phrase grammar rules also allows the text to contain a variety of topics, only some of which are pertinent to tracking time.

Description

    CROSS REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation of U.S. patent application Ser. No. 09/785,650, filed on Feb. 16, 2001, incorporated by reference herein.
  • FIELD OF THE INVENTION
  • The present invention relates to time processing and, more particularly, relates to tracking time using portable recorders and speech recognition.
  • BACKGROUND OF THE INVENTION
  • Time is a fundamental constraint for all human endeavors, and the primary source of revenue for businesses that charge clients per unit of time. Knowing where time is spent is crucial for any productivity analysis, whether on a professional or a personal level. Manual methods for capturing time, such as journals, time cards and logs, were used by professionals for many years, but in the recent past automated tools have been developed to support this process.
  • The basic data item that must be captured to begin any time analysis is a time increment, which denotes a time segment during which a particular person was dedicated to a particular task. After defining task categories for which time will be collected, a person then creates time increment records, typically by using one of two methods.
  • In the first method, time increments are recorded after the fact based on estimates. An example of this method is a project management software package that allows its users to allocate time spent on tasks over a large time interval, perhaps a whole week. For example, consultants working on a software project use the project management software at the end of the workweek to record the tasks to which they dedicated time during the previous week. In this method, there is no direct capture of the time spent and the information generated is generally imprecise.
  • In the second method, time increment data is captured directly by interacting with software running on a desktop or hand-held computer platform that allows the user to start and stop a timer at the same time that the task starts and ends. This “real-time” capture is much more accurate than the previous method, because the time increment record is captured when the task is performed and an automatic clock is used to compute the interval rather than a human estimate. However, using a computer as a capture device is inconvenient for a variety of reasons that will be discussed below.
  • Once captured with either of the above methods, time increment records can be summarized in order to generate a variety of productivity analyses or billing reports. However, when an inaccurate capture method is used, such as recording hours at the end of the week, any analysis generated is purely an approximation.
  • These prior methods have the two major problems of accuracy and ease of use. Accuracy is a huge problem for any method that relies on human memory and estimates rather than supporting automated real-time capture of time increment records. Billing generated from such data is spurious at best.
  • Ease of use issues abound in any method requiring interaction with a computer interface in order to capture time intervals. Interacting with a computer, whether a hand-held or desktop machine, generally requires a complete interruption to the task at hand and typically takes several seconds to perhaps minutes to complete. The interruption of workflow and the amount of time required becomes a barrier to using computer-based technologies. As a result, such interfaces tend to be used for minimal data capture, i.e., only capturing billable time or only some of the actual billable time. This limits the usefulness of such products for other sorts of time analysis, for example, personal productivity analysis, time analysis for non-billable workers in support roles, and, most importantly, capturing and analyzing where non-billable time is being spent by billable workers. These ease-of-use barriers also prevent professionals from capturing small increments of time spent on billable projects, such as telephone calls made from the airport or electronic mail messages composed while on a commuter train. The revenue that would otherwise be generated from such activities is therefore lost.
  • Other ease-of-use problems are slightly different depending on whether a desktop or portable computer is used. Capturing time increment records with a desktop computer is not an appropriate method for those people who need a portable device, such as people whose work requires a lot of roaming around, including consultants who have meetings or work at client sites, technical support personnel who handle service calls at various locations, detectives, sales representatives, librarians, and coaches.
  • Using a portable computer device for capture gets past the problem of portability, but introduces its own set of problems. There are significant psychological, economic, and technology adoption barrier's that limit the utility of portable or hand-held computers. Many professionals who wish to keep their interaction with computers to a minimum do not feel comfortable using such devices, and others cannot afford them. Even those people with access to portable computer devices face many ease-of-use problems. All such devices require two-handed operation and cannot be operated while walking. The entry of a time increment therefore requires the operator to completely interrupt the normal flow of his or her task.
  • Thus, what is needed is a better way of tracking time that limits the problems of accuracy and ease of use.
  • SUMMARY OF THE INVENTION
  • The present invention provides for better and simplified time tracking by using portable recorders and speech recognition. This increases accuracy, as more time should be tracked in real-time, and ease of use, as portable recorders are simple, inexpensive and convenient.
  • In general, the present invention converts speech, preferably recorded on a portable recorder, to text, analyzes the text, and determines voice commands and times when the voice commands occurred. Task names are associated with voice commands and time segments. These time segments and tasks may be packaged as time increments and stored (e.g., in a file or database) for further processing. Preferably, phrase grammar rules are used when analyzing the text, as this helps to determine voice commands. Using phrase grammar rules also allows the text to contain a variety of topics, only some of which are pertinent to tracking time.
  • A more complete understanding of the present invention, as well as further features and advantages of the present invention, will be obtained by reference to the following detailed description and drawings.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a system for tracking time using portable recorders and speech recognition, in accordance with one embodiment of the present invention;
  • FIG. 2 is a flowchart of a method for tracking time using portable recorders and speech recognition, in accordance with one embodiment of the present invention;
  • FIG. 3 is a flowchart of a method for converting speech data into time increment records, in accordance with one embodiment of the present invention;
  • FIG. 4 is a flowchart of a method for determining and creating time increments, in accordance with one embodiment of the present invention;
  • FIG. 5 illustrates a portion of speech data;
  • FIG. 6 provides an illustration of determining time increments from voice commands and times for those commands, in accordance with one embodiment of the present invention;
  • FIG. 7 illustrates phrase grammar rules in accordance with one embodiment of the present invention; and
  • FIG. 8 shows a user interface for showing, analyzing and correcting time increments, in accordance with one embodiment of the present invention.
  • DETAILED DESCRIPTION OF PREFERRED EMBODIMENTS
  • Basically, the present invention allows speech recorded on a portable recorder to be converted to time increments. Each time increment can contain a time segment and a task name. Optionally, other information, such as a matter number or comments may also be packaged into a time increment. Preferably, the present invention will act to supplement existing time and billing programs. For instance, a person could use the present invention to create time increments from speech and then import the time increments into a time and billing program. Thus, the present invention can fill a niche that currently remains unfilled.
  • Turning now to FIG. 1, this figure shows a block diagram of an exemplary system 100 for tracking time using portable recorders and speech recognition. System 100 comprises a portable speech recorder 110 and a computer system 105. Portable speech recorder 110 comprises speech data 119. Computer system 105 is any type of computer system, such as a personal computer, and it comprises processor 120 and memory 125. Speech data 119 comprises speech files 111 and 115. Speech file 111 comprises a time stamp 112, a date stamp 113, and speech 114. Similarly, speech file 115 comprises a time stamp 116, a date stamp 117, and speech 118. Memory 125 comprises speech data 130, a time-logging method 140, a speech recognition engine 145, vocabulary 150, word hypotheses 151, text output 155, time information 160, phrase grammar rules 165, voice commands 170, a time record 175, a user interface 180, a time and billing system 185, and an alternative format time record 190. Speech data 130 comprises speech files 131 and 135. Speech file 131 comprises a time stamp 132, a date stamp 133, and speech 134. Similarly, speech file 135 comprises a time stamp 136, a date stamp 137, and speech 138. Time record 175 comprises time increments 166 and 167. Computer system 105 can receive computer-readable code means from sources such as compact disk 107.
  • As is known in the art, the methods and apparatus discussed herein may be distributed as an article of manufacture, such as compact disk 107, that itself comprises a computer readable medium having computer readable code means embodied thereon. The computer readable program code means is operable, in conjunction with a computer system such as computer system 105, to carry out all or some of the steps to perform the methods or create the apparatuses discussed herein. The computer readable medium may be a recordable medium (e.g., floppy disks, hard drives, compact disks, or memory cards) or may be a transmission medium (e.g., a network comprising fiber-optics, the world-wide web, cables, or a wireless channel using time-division multiple access, code-division multiple access, or other radio-frequency channel). Any medium known or developed that can store information suitable for use with a computer system may be used. The computer-readable code means is any mechanism for allowing a computer to read instructions and data, such as magnetic variations on a magnetic medium or height variations on the surface of a compact disk.
  • Memory 125 will configure the processor 120 to implement the methods, steps, and functions disclosed herein. The memory 125 could be distributed or local and the processor could be distributed or singular. The memory 125 could be implemented as an electrical, magnetic or optical memory, or any combination of these or other types of storage devices. The term “memory” should be construed broadly enough to encompass any information able to be read from or written to an address in the addressable space accessed by processor 120. With this definition, information on a network is still within memory 125 because the processor 120 can retrieve the information from the network. It should be noted that if processor 120 is distributed, there could be multiple address spaces, one address space per processing unit that makes up processor 120.
  • Portable recorder 110 is any type of recorder that can store speech. Preferably, portable recorder 110 records a time stamp for each amount of speech that is recorded. Preferably, this time stamp indicates when the speech starts, but the time stamp could also indicate when the speech ends.
  • In FIG. 1, portable recorder 110 contains speech data 119. Speech data 119 contains a number of speech files, of which speech files 111 and 115 are shown. Speech data 119 could be stored on a tape or other memory device. Preferably, speech files contain time and date stamps along with speech. For instance, speech file 111 contains time stamp 112, date stamp 113 and speech 114, while speech file 115 contains time stamp 116, date stamp 117 and speech 118. The time and date stamps could be speech, vocalized by an operator of the portable recorder 110, of the time and date. Preferably, however, the time and date stamps are determined by a clock (not shown) of the portable recorder 110 and are placed onto the speech data 119 by portable recorder 110.
  • Speech 111 and 115 is analog or digital information of the voice of a person. Preferably, the speech 111, 115 is digital information that can be easily loaded into computer system 105. Analog speech may also be loaded into a computer system, but it is usually converted to digital speech. If speech 114 and 118 is digital speech, it could be compressed or uncompressed, although, depending on the speech recognition engine used, it may have to be decompressed to be used with a speech recognition engine.
  • In a preferred embodiment, the portable recorder 110 is a digital recorder such as the Olympus D1000 digital voice recorder. This recorder digitally stores speech information, and it can add date and time stamps to speech. The speech is stored in a format known as Digital Speech Standard (DSS), which provides for compression of the speech. The speech can be decompressed through a DSS program (not shown in FIG. 1) provided as part of a package with the D1000. Alternatively, this functionality could be built into time-logging method 140, to be discussed below.
  • A digital recorder is preferred for several reasons. First, some analog media, such as tapes, degrade fairly quickly. Moreover, even when new, tapes do not contain very good quality speech recordings. Consequently, speech recognition tends to be poor with tape recordings. Conversely, digital recorders are better at retaining a high quality of speech. This will aid in speech recognition. Another consideration is the overall quality of speech on a portable recorder. If a portable recorder is used in a quiet environment, such as an office having a door and no background noise, then even the relatively poor quality of speech recorded on a tape may be converted to text with a relatively high degree of accuracy. In general, however, a portable recorder will be used in many places where the background or ambient noise is high. The voice of the speaker in these situations will be relatively hard to determine. A digital portable recorder provides the best recording of speech under such conditions, and digital recorders can contain special circuits and software/firmware to reduce background noise. A final consideration is that a digital portable recorder can place time and date stamps on the speech data. This makes the determination of times easier because the times do not have to be converted to text, which always has a probability of error.
  • Portable recorder 110, in embodiments of the present invention that use digital portable recorders, can be any type of device that can digitally record speech and that can provide time stamps for the speech. For instance, a portable recorder 110 could be a digital personal assistant with a program to allow speech to be recorded and time stamped. As previously discussed, portable recorder 110 is preferably a digital voice portable recorder.
  • Speech data 130 is a representation of speech data 119, after speech data 119 has been loaded into computer system 105. If speech data 119 is analog speech data, the computer system 105 can digitize this data before storing it as speech data 130. If portable recorder 110 is a digital voice portable recorder, then speech data 130 will be the representation of speech data 119 that is loaded to the computer system 105 by the portable recorder 110 and any program associated with transferring the speech data 119 to speech data 130. Speech data 130 has a number of speech files 131 through 135. Speech file 131 has time stamp 132, date stamp 133 and speech 134. Preferably, the time stamp 132 and date stamp 133 are digital representations of the time and date, respectively. The time stamp 132 should correspond to time stamp 112, date stamp 133 should correspond to date stamp 113, and speech 134 should correspond to speech 114. Speech file 135 has time stamp 136, date stamp 137 and speech 138. Preferably, the time stamp 136 and date stamp 137 are digital representations of the time and date, respectively. The time stamp 136 should correspond to time stamp 116, date stamp 137 should correspond to date stamp 117, and speech 138 should correspond to speech 118.
  • If a digital voice portable recorder 110, such as a D1000 portable recorder made by Olympus, is used, the speech data 130 could be in a proprietary format. However, the time stamps 132, 136 and date stamps 133, 137 can still be accessed, deciphered and correlated with the appropriate speech 134, 138. Furthermore, as discussed below, the speech can still be accessed for conversion to text.
  • Time-logging method 140 controls the computer system 105 to convert speech in speech data 130 to time increment record 175. Optionally, time-logging method may also provide a user interface 180 (to be discussed below) and can facilitate transfer of speech data 119 into speech data 130. Time-logging method 140 can contain some, many or all of the steps necessary to perform the methods and create the apparatus or rules discussed below in reference to FIGS. 3 through 8.
  • In particular, the time-logging method 140 accesses or interfaces with the speech recognition engine 145 to cause the engine to convert speech in speech data 130 to text. As is known in the art, the speech recognition engine 145 uses a vocabulary 150 to determine appropriate word hypotheses 151. The speech recognition engine 145 selects the word hypothesis with the highest probability from the word hypotheses 151 to create text output 155. Text output 155 could contain multiple text files. For instance, there could be a text file corresponding to speech 134 and another text file corresponding to speech 138. Alternatively, there could be one large text file that contains all the text, perhaps with demarcations to indicate snippets of text and different start times.
  • Preferably, the speech recognition engine 145 also produces time information 160 that allows the text in text output 155 to be tracked relative to a starting time. As explained in more detail in reference to FIG. 5, time information 160 is beneficial to determine when a voice command was said relative to a time stamp.
  • Speech recognition engine 145 can be any of a number of speech recognition engines, such as the speech recognition engine that comes with VIAVOICE, which is a popular speech recognition program from IBM (International Business Machines, Incorporated, New Orchard Road, Armonk, N.Y., 10504). As discussed above and discussed in more detail below in reference to FIG. 6, it is preferred that the speech recognition engine used provide time information that relates the time that text occurs with a relative time of speech. The speech engine VIAVOICE makes this type of timing information available. A program written in the TCL language can be written to access the time information from VIAVOICE, and this information will be relative to the beginning of a speech file.
  • Time-logging method 140 can then compare the words in the text output 155 with phrase grammar rules 165 to determine voice commands 170. The phrase grammar rules 165 allows system 100 to specifically look for certain key words or terms. These key words and terms are the particular voice commands for which the system is trained. The voice commands are speech snippets that indicate to computer system 105 that the user wants time logging performed. For instance, the user could use the voice commands “going home,” “starting clientA,” or “ending clientA.” These are terms that the computer system 105 should find and decipher to be able to log time. The time-logging method 140 uses the voice commands 170 to create time record 175, which contains time increments 176 through 177. Each time increment, as shown in FIG. 6 below, will preferably contain a time segment and a task name. The task name will preferably be a nickname that can be expanded, if desired, by time-logging method 140. This is discussed in more detail in reference to FIG. 8.
  • User interface 180 is an optional but preferred component of time-logging method 140. Generally, user interface 180 will be part of a software package (not shown) that contains time-logging method 140, and user interface 180 may be made integral to time-logging method 140. User interface 180 allows an operator to perform analyses on the time increments 176, 177 of time record 175. Also, the user interface can allow the user to enter or change data in the time record 175. The user interface 180 may also convert the time record 175 to a form suitable to importing into a time and billing system, such as time and billing system 185. In FIG. 1, an alternative format time record 190 is shown that has been converted from time record 175 into a form suitable for importation into time and billing system 185. For instance, TIMESLIPS is a time and billing program owned by Sage U.S. Holdings, Incorporated, a wholly-owned subsidiary of The Sage Group, Plc., based in Newcastle-upon-Tyne, England. TIMESLIPS contains an importation program called TSIMPORT that can import data from a source file into a TIMESLIPS database. TSIMPORT can import tab-delimited files (where fields are separated by tabs), comma-delimited files (where fields are separated by commas), custom-delimited files (where fields are separated by delimiters specified by the user), and other types of files. The user interface 180 allows a user to convert time record 175 into one or more of these types of files. Alternatively, it is possible to create time record 175 directly in one of these types of file formats such that no conversion need take place. However, the user interface 180 is still beneficial, as it allows correcting and analyzing data in time record 175 prior to importing the time records into time and billing system 185.
  • Thus, FIG. 1 shows a system 100 that allows a person to record voice on a portable recorder and that creates a time record from spoken commands. System 100 is convenient because all a user generally has is a portable recorder into which the user dictates. The conversion from speech data to time increments and time records can occur with any computer system having the appropriate software installed. In particular, computer system 105 could be a network ed server and speech data 119 could be loaded onto the computer system 105 through a network.
  • In FIG. 1, it is also possible for speech recognition engine 145, vocabulary 150, and word hypotheses 151 to be part of portable recorder 110. This could occur, e.g., if the portable recorder is a personal digital assistant that can support speech processing. In this case, the text output 155 would be uploaded to the computer system 105 for processing. Also, time information 160 could be determined by having the computer system 105 query the speech recognition engine 145 (now on the portable recorder 110) for the time information 160 or having the portable recorder 110 upload the time information 160 to the computer system 105.
  • Referring now to FIG. 2, this figure shows a flow chart of an overall method 200 for tracking time using portable recorders and speech recognition, in accordance with one embodiment of the present invention. Method 200 is performed whenever a computer system is being prepared to convert speech into time increments. Once the computer system is prepared to convert speech into time increments, then some steps (as discussed below) of method 200 may have already been performed.
  • Method 200 begins when speech recognition and time logging software are installed on a computer system (step 210). This step generally involves using an article of manufacture, such as a compact disk, to configure a system to run the speech recognition and time logging software. Step 210 is used to prepare a computer system to recognize speech and change speech into time increments. Once the computer system is prepared, step 210 generally does not need to be performed again.
  • In step 220, a user can create time-tracking categories by using a user interface. Such time-tracking categories could include a task name (or nickname) and other information pertinent to a task, such as billable rate, comments, and billing category or matter number. Alternatively or in addition to creation, the time-tracking categories may be imported from a currently existing time and expense tracking program. Time-tracking categories will be discussed below in reference to FIG. 8.
  • In step 230, the user trains the speech recognition software on the pronunciation of voice commands. The present invention can be used with many speech recognition engines, and these engines could be solely used to convert speech to text for time tracking only. If this is the case, then the speech recognition software can have a limited vocabulary. Training the speech recognition software is always beneficial, regardless of the size of the intended vocabulary, as different people will pronounce the same words differently. Alternately, the speech recognition software could be generally trained for a variety of words, unrelated to the voice commands. It is still beneficial to specifically train the speech recognition software for voice commands, as it is unlikely that the vocabulary of the speech recognition software would contain some of the relatively specific information in a voice command, such as a task name (e.g., “clientA”). Once the speech recognition system has been trained, step 230 may not have to be performed again. This is particularly true if the speech recognition system can learn new words through feedback. For instance, step 230 could be periodically run to add new voice commands or tracking categories to the system.
  • In step 240, the user creates speech data on a portable recorder by speaking voice commands into the portable recorder. The user may also speak additional information, such as dictated letters, grocery lists or important thoughts, onto the portable recorder. The present invention can separate the voice commands from the other speech.
  • In step 250, the user loads the speech data onto the computer system. This could be performed through any mechanism known to those skilled in the art, such as a wireless infrared system, serial cables, a universal serial bus connection, or a firewire connection. Once the speech is loaded into the computer system, the speech data is converted into time increment records, which can be packaged into a time record. This is discussed in more detail in FIG. 3, below.
  • In step 270, the user has the option of loading the time increment records into a user interface for review, correction or analysis. An exemplary user interface is shown in FIG. 8. In step 280, the user has the additional option of converting, if necessary, the time records into a file suitable for importation into a time and billing software package. This allows the time increment records created with the present invention to be used with existing software.
  • Referring now to FIG. 3, this figure shows a flow chart of an exemplary method 260 for converting speech data into time increment records. Method 260 is performed whenever a user would like speech converted to time increment records. Generally, this occurs after the user speaks voice commands into a portable recorder for period of a day or several days. The user then loads the speech data containing the voice commands into a computer system, which then performs method 260.
  • Method 260 begins when the acoustic signal, which represents speech 320, is compared to predefined vocabulary words from a vocabulary such as vocabulary 150. Using a vocabulary is common in speech processing. A speech recognition engine can create recognized word hypotheses (step 330), which are essentially hypotheses, having certain probabilities, that particular words have been said. From the recognized word hypotheses, the speech recognition engine selects the most probable words. Thus, at the end of step 330, the speech recognition engine has changed speech into text.
  • In step 340, phrases are composed from words. Phrase grammar rules 165 are used during this step to help in determining phrases associated with voice commands. In step 350, the voice commands are extracted from the phrases and indexed. Extracting the voice commands could be simple, because the phrases should be highly representative of voice commands. However, there are times, as explained in more detail below in reference to FIG. 5, when additional processing might be needed to distinguish terms used in voice commands from the same terms used in normal speech. In other words, there could be times when the context in which a term is used will determine whether the term is or is not a voice command. It is possible to use a portable recorder only for tracking time and only for recording voice commands. In this situation, context should not be important.
  • The voice commands are indexed by determining the time, and also preferably the date, when the voice commands were recorded. As explained in more detail in reference to FIG. 6 below, this allows time increments to be created with time segments. In step 360, the time increments are created. If the speech is long and has many voice commands on it, the number of time increments could also be quite long. Once the time increments are created, they are written to long- or short-term memory for further processing. They may also be packaged into a time record, which can be a file having a specific, binary, or ASCII format.
  • To determine and create time increments, method 360, shown in FIGS. 4A and 4B, is used. Method 360 will be discussed and then several examples will be given to aid in understanding the method. Method 360 begins when a voice command is selected. Generally, this would be a voice command at the beginning of a list of such commands. In step 410, it is determined if the voice command contains a start marker, which is a marker before or after a task name, that indicates that a task is starting. The various start and end markers are shown in FIG. 7 and discussed in more detail below in reference to that figure. As a short introduction here, start markers can occur before or after a task name or could, optionally but not preferred, be task names. For example, in the voice commands “start clientA” and “clientA starting,” the words “start” and “starting” are start markers and “clientA” is a task name.
  • In FIG. 4, if the voice command does contain a start marker (step 410=YES), it is determined if there is an active task (step 415). By “active” it is meant that, according to the immediately preceding voice command, a task associated with the preceding voice command has been started. For instance, if the previous voice command was the statement, converted to text, “begin ClientA,” then the task associated with the task name ClientA is currently being performed If there is an active task (step 415=YES), the active task is ended (step 420) and a time increment for the task is created (step 425). If there is no active task (step 415=NO) or step 425 has been performed, a new task is activated (step 430). This means that a task name and start time are preferably kept such that it is known that this task is active.
  • When activating the new task, if there is no task name in the selected voice command, a place holder name such as “unknown” could be used. For example, if the current voice command is “start” but there is no task name, then “unknown” might be used. This could occur, for instance, in cases where the task name is converted to different text or is unintelligible. If the next voice command is “end ClientA,” then the place holder name could be changed to ClientA.
  • It should be noted that steps 420, 425 and 430 allow starting voice commands to end the previously active task. Consider the following situation: the voice command “start clientA” followed by any one of the subsequent voice commands “start clientB,” “going to lunch,” “going to meeting,” or “going home.” The task “clientA” will be an active task when the subsequent voice command is issued. The subsequent voice command actually starts a different task. The present invention will stop the task of “clientA” and start the subsequent task (“clientB,” lunch, meeting, or home/end of day). Thus, an ending voice command is not required to end the currently active task. This is in marked contrast to current time tracking systems, which do not and cannot assume that one task is ending if another task has begun. Thus, the present invention can “intelligently” determine when a currently active task should end.
  • In step 435, it is determined if there is a previous end marker (step 435). If there is no previous end marker (step 435=NO), the method continues in step 450. If there is a previous end marker (step 435=YES), the time from the previous voice command to the time of the current voice command me marked as unclassified (step 440) and a time increment for the unclassified time is created (step 443). The method then continues in step 450.
  • In step 450, it is determined if the voice command contains an end marker. End markers indicate that a task has ended and may precede a task name, follow a task name, or be stated by themselves. If the voice command does not contain an end marker, the method continues at step 490. If the voice command does contain an end marker (step 450=YES), it is determined if there is an active task (step 455). If there is no active task (step 460=NO), then a current time inclement is created (step 463) with a time segment that indicates that the previous time is unknown but that the current time is the time of the current voice command. The time inclement will be marked either as “unknown” or with the current task name from the voice command. These steps help if for instance, two end commands are spoken in a row. For example, if at 10:00 “end clientA” is spoken, and at 11:00 “end” is spoken, the time increment could indicate “??:??—11:00, Unknown.” Alternatively, the “??:??” could be replaced by “10:00,” which is the earliest time it could be.
  • If there is an active task (step 455=YES), it is determined, in step 465, if there is a current task name. If there is a current task name, it is determined if the current and the previous task names are the same (step 468). This tests for the condition where, e.g., “start clientA” occurs light before “end clientB.” To attempt to remedy this situation, the past time increment is created (step 470) and the current time inclement is created (step 475). In the previous example, if “start clientA” was said at 10:00 and “end clientB” was said at 11:00, these time increments could look like the following: “10:00—??:??, ClientA” and “??:??—11:00, ClientB,” respectively.
  • If the previous task name is equivalent to the current task name (step 468=YES), the active task is ended (step 480) and a time increment created for the active task (step 485). The method continues in step 490, which can also be reached if the voice command does not contain an end marker (step 450=NO). If there are mole voice commands to be examined (step 490=YES), the method begins again in step 405. If there are no more voice commands to be examined, the method ends in step 495.
  • Now that method 360 has been described, an example will be given and discussed using the method. In a preferred embodiment, the portable recorder is a digital voice recorder that contains time stamps along with speech. The content of the speech for this example will be “start clientA . . . end . . . start clientB start lunch . . . stop clientC,” where the ellipses indicate additional spoken words, pauses or stops After this speech is converted to text, the present invention, using grammar rules such as those discussed below in reference to FIG. 7, will determine that the voice commands are “start clientA,” “end,” “start clientB,” “start lunch,” and “stop clientC.”
  • The present invention can then examine these voice commands through method 360 For the first voice command, “start clientA,” there would be a start marker (step 410=YES) of “start,” there is no active task (step 415), so a new task is activated (step 430) with the task name of clientA. There are no previous end marker (step 435=NO), the voice command does not contain an end marker (step 450=NO), but there are voice commands (step 490=YES), so another voice command is selected (step 405). This voice command is “end.” This is an end marker, so step 410=NO and step 450=YES. There is an active task (step 455=YES), and there is no current task name in the voice command (step 465=NO). The active task is ended (step 480) and the time increment for the current task is created (step 485).
  • The method starts again on step 405. The selected voice command is “start clientB.” This voice command contains a start marker (step 410=YES), but there is no active task (step 415=NO). A new task is activated for clientB (step 430) and it is determined that there is a previous end marker (step 435=Yes). The time from when the previous voice command was recorded to the time of the current voice command is marked as unclassified (step 440) and a time increment is created (step 440) for the unclassified time.
  • Steps 450, 490 and 405 are performed. The next voice command is selected (step 405) and is the “start lunch” voice command. There is an active task (the task of clientB) in step 415, so the active task is ended (step 420) and a time increment created (step 425) for clientB A new task is activated (step 430) for the task of “lunch,” and there is no previous end marker (step 435=NO), so steps 450, 490 and 405 are performed. In step 405, the last voice command, “stop clientC,” is selected. This voice command contains an end marker, so step 450=YES. There is an active task (step 455=YES), as the task of “lunch” is still active. There is a current task name (step 465=YES), but the previous task name does not match the current task tame (step 468=NO). A past time increment is created (step 470) for the task of “lunch,” and a current time increment is created for the task of “clientC” (step 475). There are no more voice commands, so the method ends in step 495.
  • Method 360 thus attempts to provide time increments for the maximum number of voice commands. Additional steps may be added or current steps modified, by those skilled in the art, to ensure that unusual combinations of voice commands will be captured correctly by method 360.
  • Referring now to FIG. 5, this figure shows a representation of speech data 500. Speech data 500, in this example, comes from a digital voice recorder that records time and date stamps. The digital voice recorder in this example is being used to track time and as a dictation recorder.
  • Speech data 500 comprises time stamps 505, 530, 535, 555, and 580, date stamps 510, 531, 540, 560, 585, and speech 515, 532, 545, 565, and 590. When the digital voice recorder records speech data 500, it places time and date stamps whenever the recorder is started (which can include returning from a pause). For example, the person using this recorder pressed the “record” button and spoke the words “start clientA review and change letter to client.” The recorder created the time 505 and date 510 stamps. Speech 515 contains a voice command 520 and notes 525 concerning the voice command 520. Speech 532 simply contains a voice command. Speech 545 contains a voice command 550 and a note about the voice command 550. Speech 565 contains the beginning of a dictated patentability opinion letter, while speech 590 contains the end of the dictated letter and a voice command 595.
  • When the present invention examines this speech data 500, the present invention will determine that speech 520, 550 and 595 are voice commands but that speech 570 is not. The present invention can determine this because the phrase grammar rules, as shown and discussed below with reference to FIG. 7, do not allow a task name by itself to start or stop a task. In speech 570, there are no other start or end markers around the word “clientB,” and the present invention, based on phrase grammar rules, will determine that there is no voice command in speech 590.
  • Voice command 520 is associated with time stamp 505, or 9:15. In other words, the voice command 520 is assumed to have been said at about this time. Similarly, voice commands 532 and 550 are also assumed to have been said at about the times of time stamps 530 (9:25) and 535 (9:30), respectively. However, voice command 595 is said quite a bit of time after time stamp 580. To determine the time that voice command 595 was said, the present invention preferably accesses time information given by the speech recognition system. The time information should relate the text with a relative time after the speech has started. In general, speech 525, 532, 545, 565 and 590 can be considered separate files A speech recognition engine should treat each separately and should separately track time for each. The time information should start for each at the beginning of the speech. For instance, relative time 593, determined by accessing the time provided by the speech recognition engine, indicates that 4 minutes and 55 seconds have elapsed since the beginning of speech 590. This relative time 593 can be added to the time in time stamp 580 to determine that the voice command 595 was said at about 12:15 Thus, the relative time information from a speech recognition engine can be helpful to determine when a voice command was issued.
  • It should be noted that this type of relative time information, produced by a speech recognition engine, should not be needed if a digital portable recorder is used solely to track time, if the person stops and starts a digital recorder immediately prior to issuing a voice command, or if an analog portable recorder is used and the person states the time along with a voice command.
  • Referring now to FIG. 6, this figure shows a table 600 that contains voice command information 610 and time increment information 650. Voice command information 610 is compiled after the speech has been converted to text and phrase grammar rules have been used to determine voice commands. The content of the speech, which has been changed into voice command information, in FIG. 6 is most of the speech shown in FIG. 5. From the voice commands, the time increment information can be determined. For instance, method 360 of FIG. 4 could be used to determine the time increment information 650.
  • In FIG. 6, voice command information 610 comprises voice command entries 615, 620, 625, 630, and 635. Time increment information 650 comprises time increments 655, 660, 665, 670, and 675. Each voice command entry comprises a time 680 associated with a voice command 685. The present invention creates time increment information 650 from the voice command information 610. Many of the entries in the example in FIG. 6 have also been discussed previously in reference to FIG. 4.
  • Time increment 655 is created from voice command entries 615 and 620. Time increment 660 is created from voice command entries 625 and 620. Time increment 665 is created from voice command entries 630 and 625. It is assumed that starting a current task prior to ending a previous task will stop the previous task and start the current task. If desired, it is possible that each task must be stopped prior to starting another task. In this situation, the voice command entry 630 could cause an error time inclement. Time increments 670 and 675 are created from voice command entries 630 and 635. There is an error because there is no indication as to when lunch stopped and when the work for clientC began.
  • It should be noted that, if desired, silence (or the lack of an ending voice command) may be an indication that the current task is to end Fox instance, if quitting time is 6 pm, the system could record the ending time as 6 pm even though no ending voice command was given Alternatively, the system could indicate 6 pm as the ending time but indicate a possible error in the time increment. Such an error could be indicated as “6:00?” or “6:00!”
  • Referring now to FIG. 7, this figure shows a set of phrase grammar rules 165 that are used to determine voice commands from a text file. The rules shown in FIG. 7 are only examples of possible phrase grammar rules, and the actual set of rules used in an embodiment will probably include more rules than those shown, could include additional lexical terms, and could be differently written. There are a variety of ways of writing rules suitable for use with the present invention, and the rules included herein are merely examples of one way to do this. Phrase grammar rules 165 contain indications of tasks and their associated task names in the <task> category. The symbol “|” between each task name (and as used throughout FIG. 7) indicates “or,” while “+” indicates “and.” The rules 165 contain start markers, which indicate that a task is beginning. The start markers can be said before a task name, which is the <start-premarker>, or said after a task name, which is the <start-postmarker>. The rules 165 also contain end markers, which indicate that a task is ending. The end markers can be said before a task name, which is the <end-premarker>, said after a task name, which is the <end-postmarker>, or said alone, which is the <end-maker>. Optionally, the present invention can also use matter numbers, which are task-specific numbers that track more specific tasks. For instance, a matter number of 900 might indicate patentability opinions, which are tasks more specific than just general patent tasks. In the example of FIG. 7, approved matter numbers are in the <matter-number> category. Additionally, client-specific tasks may also be tracked. These tasks may be tracked by number or words A client-specific task might be, e.g., “clientA 9765,” where “9765” indicates a different task than, say, “9764.” The task “9765” might correspond to “trademark application,” while “9764” might correspond to “patent litigation.”
  • Phrases are series of words that indicate that a task has started or ended. For instance, one of the allowed phrases is indicated in <phrase> as “<start-premarker> or <end-premarker> and <task>.” Using this phrase definition, approved phrases are “start projectA,” “end lunch” or “quitting meeting.” Then the system can determine whether the phrase is starting or stopping the current task by determining if a start or end, respectively, marker is used.
  • The phrases in phrase grammar rules 165 are thus specific to the types of words used to start and end tasks. The list of phrase grammar rules 165 may be expanded at any time, and any new words added to the rules 165 may be trained in a speech recognition engine to provide for higher possibility that the new words will be correctly converted to speech.
  • It should be noted that a voice command is a spoken series of words that awe intended to end or start tasks. The voice command may also be a text representation of such spoken words. The phrases in phrase grammar rules 165 are, on the other hand, rules that define what words can be put together (and in what order) to create valid voice commands, but the phrase grammar rules are not the voice commands themselves.
  • Preferably, an initial set of predefined phrase grammar rules will be included with the system, and interface 180 will allow the user to modify, delete, or add new rules.
  • Turning now to FIG. 8, this figure shows an exemplary user interface 800 that allows a person to view, analyze, modify, export, create, and remove time increments. The user interface 800 can be created through a number of graphical user interfaces, as is known in the art. User interface 800 has a file menu (File, Consultant, Project, Time Records, Billing), a time record, billable activities, and non-billable activities. The time record shows the current time increments that are being analyzed or viewed. The “nickname” is a task name that indicates a task. The billable activities section shows a nickname, description and billing late. The non-billable activities section shows a nickname and a description. The nickname, description and billing late (or non-billing status) can be used to analyze the time inclement data and/or could be output to a file for subsequent loading into time and billing software such as TIMESLIPS. The description can be further improved to add additional information. For instance, a nickname of “kraft900” could have a description of Kraft, matter 900, patentability opinion. This additional data could be output to a file for further processing by time and billing software. A user can freely add and change the interface, and can use the interface to output files suitable for loading into a time and billing program. For example, under “time records,” an option might be “export,” which allows the further selection of different types of exportable file formats and which data, if any, from billable and non-billable activities sections will supplement or replace the time increments.
  • Thus, what has been shown is a time tacking system that can convert speech on portable recorders to time increments that can then be analyzed, changed or viewed. The time tracking system can create an output file having time increments that allow a user to import the file into a separate time and expense billing program.
  • It is to be understood that the embodiments and variations shown and described herein are merely illustrative of the principles of this invention and that various modifications may be implemented by those skilled in the art without departing from the scope and spirit of the invention. For instance, notes (such as note 525 in FIG. 5) could be added to time increments by the present invention, through expanded phrase grammar rules.

Claims (17)

1. A method, performed on a computer system, for tracking time using speech recognition, the method comprising the steps of:
accessing speech data;
recognizing at least two voice commands from the speech data, each voice command occurring at a different time;
determining a first time associated with a speaking of a first of the voice commands, wherein said first voice command identifies a start of a time interval;
determining a second time associated with a speaking of a second of the voice commands, wherein said second voice command identifies an end of said time interval; and
storing data identifying said time interval and data identifying one or more of said first voice command and second voice command, wherein
the speech data comprises a time stamp;
the step of determining a first time comprises:
determining an offset time between the time stamp and a time when the first voice command is spoken; and
determining the first time through reference to the time stamp and the offset time.
2. The method of claim 1, wherein:
the speech data comprises a time stamp;
the step of determining a first time comprises:
determining an offset time between the time stamp and a time when the first voice command is spoken; and
determining the first time through reference to the time stamp and the offset time; and
the step of determining a second time comprises:
determining a second offset time between the time stamp and a time when the second voice command is spoken; and
determining the second time through reference to the time stamp and the second offset time.
3. The method of claim 2, wherein:
the step of determining the first time through reference to the time stamp and the offset time comprises the step of adding the offset time to the time stamp to determine the first time; and
the step of determining the second time trough reference to the time stamp and the second offset time comprises the step of adding the second offset time to the time stamp to determine the second time.
4. The method of claim 1, wherein:
the speech data comprises first and second time stamps;
the step of determining a first time comprises:
determining a first offset time between the first time stamp and a time when the first voice command is spoken; and
determining the first time through reference to the first time stamp and the first offset time; and
the step of determining a second time comprises:
determining a second offset time between the second time stamp and a time when the second voice command is spoken; and
determining the second time through reference to the second time stamp and the second offset time.
5. The method of claim 1, further comprising the steps of:
recording speech onto a portable recorder; and
loading the speech data from the portable recorder to the computer system, the speech data comprising the speech and a plurality of time stamps.
6. The method of claim 1, further comprising the step of:
determining at least one task name from the text of the at least two voice commands.
7. The method of claim 6, wherein the step of determining at least one task name comprises finding the at least one task name in the text.
8. The method of claim 6, wherein the step of determining at least one task name comprises associating at least one task name to said time interval between the first and second times, wherein the at least one task name is not in the text.
9. The method of claim 6, wherein the at least one task name comprises two task names, a first task name associated with a first of the voice commands and a second task name associated with a second of the voice commands, wherein the first and second voice commands occur adjacent to each other in time, wherein the first and second task name are different, and wherein the second voice command is assumed to end a first task corresponding to the first task name and start a second task corresponding to the second task name.
10. The method of claim 6, further comprising the step of packaging the first time, second time, and one task name from the at least one task name into a time increment.
11. The method of claim 6, wherein the at least two voice commands comprises a plurality of voice commands, wherein the at least one task name comprises a plurality of task names, and wherein the method further comprises the steps of:
determining an additional plurality of voice command times, each of the voice command times associated with one of the plurality of additional voice command times;
converting each of the plurality of voice commands to text;
determining a plurality of task names from the text;
associating a task name with two of the first time, second time, or additional plurality of voice command times;
creating a plurality of time increments, each time increment comprising two times of the first time, second time, or additional plurality of voice command times and a task name; and
storing the plurality of time increments.
12. The method of claim 6, further comprising the step of determining text versions of the at least two voice commands by comparing words in the text with phrase grammar rules.
13. A system for tracking time using speech recognition, the system comprising:
a computer system comprising:
a memory that stores computer-readable code; and
a processor operatively coupled to the memory, the processor configured to implement the computer-readable code, the computer-readable code configured to:
access speech data;
recognize at least two voice commands from the speech data, each voice command occurring at a different time;
determine a first time associated with a speaking of a first of the voice commands, wherein said first voice command identifies a start of a time interval;
determine a second time associated with a speaking of a second of the voice commands, wherein said second voice command identifies an end of a time interval;
convert each of the at least two voice commands to text;
determine text versions of the at least two voice commands by comparing words in the text with phrase grammar rules; and
storing data identifying said time interval and data identifying one or more of said first voice command and second voice command, wherein
the speech data comprises a time stamp;
the step of determining a first time comprises:
determining an offset time between the time stamp and a time when the first voice command is spoken; and
determining the first time through reference to the time stamp and the offset time.
14. The system of claim 13, wherein the speech data comprises a time stamp, and wherein the computer-readable code is further configured to:
when determining a first time:
determining an offset time between the time stamp and a time when the first voice command is spoken; and
determining the first time through reference to the time stamp and the offset time; and
when determining a second time:
determining a second offset time between the time stamp and a time when the second voice command is spoken; and
determining the second time through reference to the time stamp and the second offset time.
15. The system of claim 13, wherein the computer-readable code is further configured to:
store one or more time increments comprised of the first time, the second time, and at least one text version of the at least two voice commands; and
place the time increments into a file having a format suitable for importing into a time and billing program.
16. The system of claim 13, wherein the system further comprises a digital personal recorder and wherein the computer-readable code is further configured to receive the speech data from the digital personal recorder.
17. An article of manufacture comprising:
a computer readable medium having computer readable code means embodied thereon, the computer readable program code means comprising:
a step to access speech data;
a step to recognize at least two voice commands from the speech data, each voice command occurring at a different time;
a step to determine a first time associated with a speaking of a first of the voice commands, wherein said first voice command identifies a start of a time interval;
a step to determine a second time associated with a speaking of a second of the voice commands, wherein said second voice command identifies an end of a time interval;
a step to convert each of the at least two voice commands to text;
a step to determine text versions of the at least two voice commands by comparing words in the text with phrase grammar rules; and
storing data identifying said time interval and data identifying one or mole of said first voice command and second voice command.
the speech data comprises a time stamp;
the step of determining a first time comprises:
determining an offset time between the time stamp and a time when the first voice command is spoken; and
determining the first time through reference to the time stamp and the offset time.
US11/566,802 2001-02-16 2006-12-05 Tracking Time Using Portable Recorders and Speech Recognition Abandoned US20070100634A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US11/566,802 US20070100634A1 (en) 2001-02-16 2006-12-05 Tracking Time Using Portable Recorders and Speech Recognition
US12/180,797 US7664638B2 (en) 2001-02-16 2008-07-28 Tracking time using portable recorders and speech recognition

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US09/785,650 US7171365B2 (en) 2001-02-16 2001-02-16 Tracking time using portable recorders and speech recognition
US11/566,802 US20070100634A1 (en) 2001-02-16 2006-12-05 Tracking Time Using Portable Recorders and Speech Recognition

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US09/785,650 Continuation US7171365B2 (en) 2001-02-16 2001-02-16 Tracking time using portable recorders and speech recognition

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US12/180,797 Continuation US7664638B2 (en) 2001-02-16 2008-07-28 Tracking time using portable recorders and speech recognition

Publications (1)

Publication Number Publication Date
US20070100634A1 true US20070100634A1 (en) 2007-05-03

Family

ID=25136184

Family Applications (3)

Application Number Title Priority Date Filing Date
US09/785,650 Expired - Lifetime US7171365B2 (en) 2001-02-16 2001-02-16 Tracking time using portable recorders and speech recognition
US11/566,802 Abandoned US20070100634A1 (en) 2001-02-16 2006-12-05 Tracking Time Using Portable Recorders and Speech Recognition
US12/180,797 Expired - Fee Related US7664638B2 (en) 2001-02-16 2008-07-28 Tracking time using portable recorders and speech recognition

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US09/785,650 Expired - Lifetime US7171365B2 (en) 2001-02-16 2001-02-16 Tracking time using portable recorders and speech recognition

Family Applications After (1)

Application Number Title Priority Date Filing Date
US12/180,797 Expired - Fee Related US7664638B2 (en) 2001-02-16 2008-07-28 Tracking time using portable recorders and speech recognition

Country Status (1)

Country Link
US (3) US7171365B2 (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100010814A1 (en) * 2008-07-08 2010-01-14 International Business Machines Corporation Enhancing media playback with speech recognition
US8719121B1 (en) * 2008-01-15 2014-05-06 David Leason System and method for automated construction of time records based on electronic messages
US20180034961A1 (en) * 2014-02-28 2018-02-01 Ultratec, Inc. Semiautomated Relay Method and Apparatus
US10475464B2 (en) * 2012-07-03 2019-11-12 Samsung Electronics Co., Ltd Method and apparatus for connecting service between user devices using voice
US12035070B2 (en) 2020-02-21 2024-07-09 Ultratec, Inc. Caption modification and augmentation systems and methods for use by hearing assisted user
US12137183B2 (en) 2014-02-28 2024-11-05 Ultratec, Inc. Semiautomated relay method and apparatus
US12136425B2 (en) 2014-02-28 2024-11-05 Ultratec, Inc. Semiautomated relay method and apparatus

Families Citing this family (223)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8645137B2 (en) 2000-03-16 2014-02-04 Apple Inc. Fast, language-independent method for user authentication by voice
US7171365B2 (en) * 2001-02-16 2007-01-30 International Business Machines Corporation Tracking time using portable recorders and speech recognition
GB2383733B (en) * 2001-12-28 2005-08-03 Nokia Corp A communication log for an electronic device
US7149693B2 (en) * 2003-07-31 2006-12-12 Sony Corporation Automated digital voice recorder to personal information manager synchronization
KR100554442B1 (en) * 2003-10-06 2006-02-22 주식회사 팬택앤큐리텔 Mobile Communication Terminal with Voice Recognition function, Phoneme Modeling Method and Voice Recognition Method for the same
US7555533B2 (en) * 2003-10-15 2009-06-30 Harman Becker Automotive Systems Gmbh System for communicating information from a server via a mobile communication device
EP1555652B1 (en) * 2004-01-19 2007-11-14 Harman Becker Automotive Systems GmbH Activation of a speech dialogue system
DE602004017955D1 (en) * 2004-01-29 2009-01-08 Daimler Ag Method and system for voice dialogue interface
DE602004014893D1 (en) 2004-01-29 2008-08-21 Daimler Ag Multimodal data entry
EP1562180B1 (en) * 2004-02-06 2015-04-01 Nuance Communications, Inc. Speech dialogue system and method for controlling an electronic device
US7756738B2 (en) * 2004-08-12 2010-07-13 Lumen Patent Firm Professional service management using project-based point system
US7564959B2 (en) * 2004-11-15 2009-07-21 Microsoft Corporation Billable activity processing
US8033831B2 (en) * 2004-11-22 2011-10-11 Bravobrava L.L.C. System and method for programmatically evaluating and aiding a person learning a new language
US8272874B2 (en) * 2004-11-22 2012-09-25 Bravobrava L.L.C. System and method for assisting language learning
US8221126B2 (en) * 2004-11-22 2012-07-17 Bravobrava L.L.C. System and method for performing programmatic language learning tests and evaluations
US20060143272A1 (en) * 2004-12-27 2006-06-29 Alcatel Electronic communications content elaboration assistant
US7004389B1 (en) 2005-01-13 2006-02-28 Biopay, Llc System and method for tracking a mobile worker
US8677377B2 (en) 2005-09-08 2014-03-18 Apple Inc. Method and apparatus for building an intelligent automated assistant
US9318108B2 (en) 2010-01-18 2016-04-19 Apple Inc. Intelligent automated assistant
US7881990B2 (en) * 2006-11-30 2011-02-01 Intuit Inc. Automatic time tracking based on user interface events
US8977255B2 (en) 2007-04-03 2015-03-10 Apple Inc. Method and system for operating a multi-function portable electronic device using voice-activation
US10002189B2 (en) 2007-12-20 2018-06-19 Apple Inc. Method and apparatus for searching using an active ontology
US9330720B2 (en) 2008-01-03 2016-05-03 Apple Inc. Methods and apparatus for altering audio output signals
US20100293021A1 (en) * 2009-01-23 2010-11-18 Intranet Productivity Solutions, Ltd. Method and system for task tracking and allocation
US8996376B2 (en) 2008-04-05 2015-03-31 Apple Inc. Intelligent text-to-speech conversion
US10496753B2 (en) 2010-01-18 2019-12-03 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US20100030549A1 (en) 2008-07-31 2010-02-04 Lee Michael M Mobile device having human language translation capability with positional feedback
US8676904B2 (en) 2008-10-02 2014-03-18 Apple Inc. Electronic devices with voice command and contextual data processing capabilities
WO2010067118A1 (en) 2008-12-11 2010-06-17 Novauris Technologies Limited Speech recognition involving a mobile device
US9858925B2 (en) 2009-06-05 2018-01-02 Apple Inc. Using context information to facilitate processing of commands in a virtual assistant
US10241644B2 (en) 2011-06-03 2019-03-26 Apple Inc. Actionable reminder entries
US20120309363A1 (en) 2011-06-03 2012-12-06 Apple Inc. Triggering notifications associated with tasks items that represent tasks to perform
US10241752B2 (en) 2011-09-30 2019-03-26 Apple Inc. Interface for a virtual digital assistant
US9431006B2 (en) 2009-07-02 2016-08-30 Apple Inc. Methods and apparatuses for automatic speech recognition
US10553209B2 (en) 2010-01-18 2020-02-04 Apple Inc. Systems and methods for hands-free notification summaries
US10705794B2 (en) 2010-01-18 2020-07-07 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US10679605B2 (en) 2010-01-18 2020-06-09 Apple Inc. Hands-free list-reading by intelligent automated assistant
US10276170B2 (en) 2010-01-18 2019-04-30 Apple Inc. Intelligent automated assistant
DE112011100329T5 (en) 2010-01-25 2012-10-31 Andrew Peter Nelson Jerram Apparatus, methods and systems for a digital conversation management platform
US8682667B2 (en) 2010-02-25 2014-03-25 Apple Inc. User profiling for selecting user specific voice input processing information
US20110218802A1 (en) * 2010-03-08 2011-09-08 Shlomi Hai Bouganim Continuous Speech Recognition
US10762293B2 (en) 2010-12-22 2020-09-01 Apple Inc. Using parts-of-speech tagging and named entity recognition for spelling correction
US9262612B2 (en) 2011-03-21 2016-02-16 Apple Inc. Device access using voice authentication
US10057736B2 (en) 2011-06-03 2018-08-21 Apple Inc. Active transport based notifications
US10672399B2 (en) 2011-06-03 2020-06-02 Apple Inc. Switching between text data and audio data based on a mapping
US8994660B2 (en) 2011-08-29 2015-03-31 Apple Inc. Text correction processing
US10134385B2 (en) 2012-03-02 2018-11-20 Apple Inc. Systems and methods for name pronunciation
US8798996B2 (en) * 2012-03-05 2014-08-05 Coupons.Com Incorporated Splitting term lists recognized from speech
US9483461B2 (en) 2012-03-06 2016-11-01 Apple Inc. Handling speech synthesis of content for multiple languages
US9280610B2 (en) 2012-05-14 2016-03-08 Apple Inc. Crowd sourcing information to fulfill user requests
US10417037B2 (en) 2012-05-15 2019-09-17 Apple Inc. Systems and methods for integrating third party services with a digital assistant
KR20130133629A (en) 2012-05-29 2013-12-09 삼성전자주식회사 Method and apparatus for executing voice command in electronic device
US9721563B2 (en) 2012-06-08 2017-08-01 Apple Inc. Name recognition system
US9495129B2 (en) 2012-06-29 2016-11-15 Apple Inc. Device, method, and user interface for voice-activated navigation and browsing of a document
US9576574B2 (en) 2012-09-10 2017-02-21 Apple Inc. Context-sensitive handling of interruptions by intelligent digital assistant
US9547647B2 (en) 2012-09-19 2017-01-17 Apple Inc. Voice-based media searching
CN113744733B (en) 2013-02-07 2022-10-25 苹果公司 Voice trigger of digital assistant
US9549046B1 (en) * 2013-02-20 2017-01-17 The Boeing Company Methods and systems for logging time with a mobile computing device
US9733821B2 (en) 2013-03-14 2017-08-15 Apple Inc. Voice control to diagnose inadvertent activation of accessibility features
US9977779B2 (en) 2013-03-14 2018-05-22 Apple Inc. Automatic supplementation of word correction dictionaries
US9368114B2 (en) 2013-03-14 2016-06-14 Apple Inc. Context-sensitive handling of interruptions
US10572476B2 (en) 2013-03-14 2020-02-25 Apple Inc. Refining a search based on schedule items
US10642574B2 (en) 2013-03-14 2020-05-05 Apple Inc. Device, method, and graphical user interface for outputting captions
US10652394B2 (en) 2013-03-14 2020-05-12 Apple Inc. System and method for processing voicemail
AU2014233517B2 (en) 2013-03-15 2017-05-25 Apple Inc. Training an at least partial voice command system
WO2014144579A1 (en) 2013-03-15 2014-09-18 Apple Inc. System and method for updating an adaptive speech recognition model
AU2014251347B2 (en) 2013-03-15 2017-05-18 Apple Inc. Context-sensitive handling of interruptions
US10748529B1 (en) 2013-03-15 2020-08-18 Apple Inc. Voice activated device for use with a voice-based digital assistant
KR102014665B1 (en) 2013-03-15 2019-08-26 애플 인크. User training by intelligent digital assistant
WO2014197336A1 (en) 2013-06-07 2014-12-11 Apple Inc. System and method for detecting errors in interactions with a voice-based digital assistant
US9582608B2 (en) 2013-06-07 2017-02-28 Apple Inc. Unified ranking with entropy-weighted information for phrase-based semantic auto-completion
WO2014197334A2 (en) 2013-06-07 2014-12-11 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
WO2014197335A1 (en) 2013-06-08 2014-12-11 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
US10176167B2 (en) 2013-06-09 2019-01-08 Apple Inc. System and method for inferring user intent from speech inputs
AU2014278592B2 (en) 2013-06-09 2017-09-07 Apple Inc. Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant
JP2016521948A (en) 2013-06-13 2016-07-25 アップル インコーポレイテッド System and method for emergency calls initiated by voice command
CN105453026A (en) 2013-08-06 2016-03-30 苹果公司 Auto-activating smart responses based on activities from remote devices
US10296160B2 (en) 2013-12-06 2019-05-21 Apple Inc. Method for extracting salient dialog usage from live data
US9620105B2 (en) 2014-05-15 2017-04-11 Apple Inc. Analyzing audio input for efficient speech and music recognition
US10592095B2 (en) 2014-05-23 2020-03-17 Apple Inc. Instantaneous speaking of content on touch devices
US9502031B2 (en) 2014-05-27 2016-11-22 Apple Inc. Method for supporting dynamic grammars in WFST-based ASR
US10078631B2 (en) 2014-05-30 2018-09-18 Apple Inc. Entropy-guided text prediction using combined word and character n-gram language models
US9760559B2 (en) 2014-05-30 2017-09-12 Apple Inc. Predictive text input
US9842101B2 (en) 2014-05-30 2017-12-12 Apple Inc. Predictive conversion of language input
US9785630B2 (en) 2014-05-30 2017-10-10 Apple Inc. Text prediction using combined word N-gram and unigram language models
US9734193B2 (en) 2014-05-30 2017-08-15 Apple Inc. Determining domain salience ranking from ambiguous words in natural speech
US10289433B2 (en) 2014-05-30 2019-05-14 Apple Inc. Domain specific language for encoding assistant dialog
US10170123B2 (en) 2014-05-30 2019-01-01 Apple Inc. Intelligent assistant for home automation
CN110797019B (en) 2014-05-30 2023-08-29 苹果公司 Multi-command single speech input method
US9430463B2 (en) 2014-05-30 2016-08-30 Apple Inc. Exemplar-based natural language processing
US9715875B2 (en) 2014-05-30 2017-07-25 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US9633004B2 (en) 2014-05-30 2017-04-25 Apple Inc. Better resolution when referencing to concepts
US9338493B2 (en) 2014-06-30 2016-05-10 Apple Inc. Intelligent automated assistant for TV user interactions
US10659851B2 (en) 2014-06-30 2020-05-19 Apple Inc. Real-time digital assistant knowledge updates
US10446141B2 (en) 2014-08-28 2019-10-15 Apple Inc. Automatic speech recognition based on user feedback
US9953646B2 (en) 2014-09-02 2018-04-24 Belleau Technologies Method and system for dynamic speech recognition and tracking of prewritten script
US9818400B2 (en) 2014-09-11 2017-11-14 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US10789041B2 (en) 2014-09-12 2020-09-29 Apple Inc. Dynamic thresholds for always listening speech trigger
US9646609B2 (en) 2014-09-30 2017-05-09 Apple Inc. Caching apparatus for serving phonetic pronunciations
US9668121B2 (en) 2014-09-30 2017-05-30 Apple Inc. Social reminders
US10127911B2 (en) 2014-09-30 2018-11-13 Apple Inc. Speaker identification and unsupervised speaker adaptation techniques
US10074360B2 (en) 2014-09-30 2018-09-11 Apple Inc. Providing an indication of the suitability of speech recognition
US9886432B2 (en) 2014-09-30 2018-02-06 Apple Inc. Parsimonious handling of word inflection via categorical stem + suffix N-gram language models
US10552013B2 (en) 2014-12-02 2020-02-04 Apple Inc. Data detection
US9711141B2 (en) 2014-12-09 2017-07-18 Apple Inc. Disambiguating heteronyms in speech synthesis
US10152299B2 (en) 2015-03-06 2018-12-11 Apple Inc. Reducing response latency of intelligent automated assistants
US9865280B2 (en) 2015-03-06 2018-01-09 Apple Inc. Structured dictation using intelligent automated assistants
US10567477B2 (en) 2015-03-08 2020-02-18 Apple Inc. Virtual assistant continuity
US9886953B2 (en) 2015-03-08 2018-02-06 Apple Inc. Virtual assistant activation
US9721566B2 (en) 2015-03-08 2017-08-01 Apple Inc. Competing devices responding to voice triggers
US9899019B2 (en) 2015-03-18 2018-02-20 Apple Inc. Systems and methods for structured stem and suffix language models
US9842105B2 (en) 2015-04-16 2017-12-12 Apple Inc. Parsimonious continuous-space phrase representations for natural language processing
US10460227B2 (en) 2015-05-15 2019-10-29 Apple Inc. Virtual assistant in a communication session
US10083688B2 (en) 2015-05-27 2018-09-25 Apple Inc. Device voice control for selecting a displayed affordance
US10127220B2 (en) 2015-06-04 2018-11-13 Apple Inc. Language identification from short strings
US9578173B2 (en) 2015-06-05 2017-02-21 Apple Inc. Virtual assistant aided communication with 3rd party service in a communication session
US10101822B2 (en) 2015-06-05 2018-10-16 Apple Inc. Language input correction
US11025565B2 (en) 2015-06-07 2021-06-01 Apple Inc. Personalized prediction of responses for instant messaging
US10255907B2 (en) 2015-06-07 2019-04-09 Apple Inc. Automatic accent detection using acoustic models
US10186254B2 (en) 2015-06-07 2019-01-22 Apple Inc. Context-based endpoint detection
US20160378747A1 (en) 2015-06-29 2016-12-29 Apple Inc. Virtual assistant for media playback
US10747498B2 (en) 2015-09-08 2020-08-18 Apple Inc. Zero latency digital assistant
US10671428B2 (en) 2015-09-08 2020-06-02 Apple Inc. Distributed personal assistant
US9697820B2 (en) 2015-09-24 2017-07-04 Apple Inc. Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks
US11010550B2 (en) 2015-09-29 2021-05-18 Apple Inc. Unified language modeling framework for word prediction, auto-completion and auto-correction
US10366158B2 (en) 2015-09-29 2019-07-30 Apple Inc. Efficient word encoding for recurrent neural network language models
US11587559B2 (en) 2015-09-30 2023-02-21 Apple Inc. Intelligent device identification
US10691473B2 (en) 2015-11-06 2020-06-23 Apple Inc. Intelligent automated assistant in a messaging environment
US10049668B2 (en) 2015-12-02 2018-08-14 Apple Inc. Applying neural network language models to weighted finite state transducers for automatic speech recognition
US10223066B2 (en) 2015-12-23 2019-03-05 Apple Inc. Proactive assistance based on dialog communication between devices
US11423348B2 (en) * 2016-01-11 2022-08-23 Hand Held Products, Inc. System and method for assessing worker performance
US10446143B2 (en) 2016-03-14 2019-10-15 Apple Inc. Identification of voice inputs providing credentials
US9934775B2 (en) 2016-05-26 2018-04-03 Apple Inc. Unit-selection text-to-speech synthesis based on predicted concatenation parameters
US9972304B2 (en) 2016-06-03 2018-05-15 Apple Inc. Privacy preserving distributed evaluation framework for embedded personalized systems
US11227589B2 (en) 2016-06-06 2022-01-18 Apple Inc. Intelligent list reading
US10249300B2 (en) 2016-06-06 2019-04-02 Apple Inc. Intelligent list reading
US10049663B2 (en) 2016-06-08 2018-08-14 Apple, Inc. Intelligent automated assistant for media exploration
DK179309B1 (en) 2016-06-09 2018-04-23 Apple Inc Intelligent automated assistant in a home environment
US10509862B2 (en) 2016-06-10 2019-12-17 Apple Inc. Dynamic phrase expansion of language input
US10586535B2 (en) 2016-06-10 2020-03-10 Apple Inc. Intelligent digital assistant in a multi-tasking environment
US10490187B2 (en) 2016-06-10 2019-11-26 Apple Inc. Digital assistant providing automated status report
US10192552B2 (en) 2016-06-10 2019-01-29 Apple Inc. Digital assistant providing whispered speech
US10067938B2 (en) 2016-06-10 2018-09-04 Apple Inc. Multilingual word prediction
DK201670540A1 (en) 2016-06-11 2018-01-08 Apple Inc Application integration with a digital assistant
US10621581B2 (en) 2016-06-11 2020-04-14 Apple Inc. User interface for transactions
DK179343B1 (en) 2016-06-11 2018-05-14 Apple Inc Intelligent task discovery
DK179415B1 (en) 2016-06-11 2018-06-14 Apple Inc Intelligent device arbitration and control
DK179049B1 (en) 2016-06-11 2017-09-18 Apple Inc Data driven natural language event detection and classification
CN109313759B (en) 2016-06-11 2022-04-26 苹果公司 User interface for transactions
US20180068313A1 (en) 2016-09-06 2018-03-08 Apple Inc. User interfaces for stored-value accounts
US10474753B2 (en) 2016-09-07 2019-11-12 Apple Inc. Language identification using recurrent neural networks
US10043516B2 (en) 2016-09-23 2018-08-07 Apple Inc. Intelligent automated assistant
US11281993B2 (en) 2016-12-05 2022-03-22 Apple Inc. Model and ensemble compression for metric learning
US10593346B2 (en) 2016-12-22 2020-03-17 Apple Inc. Rank-reduced token representation for automatic speech recognition
US11204787B2 (en) 2017-01-09 2021-12-21 Apple Inc. Application integration with a digital assistant
US11010601B2 (en) 2017-02-14 2021-05-18 Microsoft Technology Licensing, Llc Intelligent assistant device communicating non-verbal cues
US11100384B2 (en) 2017-02-14 2021-08-24 Microsoft Technology Licensing, Llc Intelligent device user interactions
US10467509B2 (en) 2017-02-14 2019-11-05 Microsoft Technology Licensing, Llc Computationally-efficient human-identifying smart assistant computer
DK201770383A1 (en) 2017-05-09 2018-12-14 Apple Inc. User interface for correcting recognition errors
US10417266B2 (en) 2017-05-09 2019-09-17 Apple Inc. Context-aware ranking of intelligent response suggestions
US10726832B2 (en) 2017-05-11 2020-07-28 Apple Inc. Maintaining privacy of personal information
US10395654B2 (en) 2017-05-11 2019-08-27 Apple Inc. Text normalization based on a data-driven learning network
DK201770439A1 (en) 2017-05-11 2018-12-13 Apple Inc. Offline personal assistant
DK201770429A1 (en) 2017-05-12 2018-12-14 Apple Inc. Low-latency intelligent automated assistant
DK179496B1 (en) 2017-05-12 2019-01-15 Apple Inc. USER-SPECIFIC Acoustic Models
US11301477B2 (en) 2017-05-12 2022-04-12 Apple Inc. Feedback analysis of a digital assistant
DK179745B1 (en) 2017-05-12 2019-05-01 Apple Inc. SYNCHRONIZATION AND TASK DELEGATION OF A DIGITAL ASSISTANT
DK201770431A1 (en) 2017-05-15 2018-12-20 Apple Inc. Optimizing dialogue policy decisions for digital assistants using implicit feedback
DK201770432A1 (en) 2017-05-15 2018-12-21 Apple Inc. Hierarchical belief states for digital assistants
US10303715B2 (en) 2017-05-16 2019-05-28 Apple Inc. Intelligent automated assistant for media exploration
US10403278B2 (en) 2017-05-16 2019-09-03 Apple Inc. Methods and systems for phonetic matching in digital assistant services
DK179549B1 (en) 2017-05-16 2019-02-12 Apple Inc. Far-field extension for digital assistant services
US11221744B2 (en) 2017-05-16 2022-01-11 Apple Inc. User interfaces for peer-to-peer transfers
KR102495947B1 (en) 2017-05-16 2023-02-06 애플 인크. User interfaces for peer-to-peer transfers
US10311144B2 (en) 2017-05-16 2019-06-04 Apple Inc. Emoji word sense disambiguation
US20180336892A1 (en) 2017-05-16 2018-11-22 Apple Inc. Detecting a trigger of a digital assistant
US10657328B2 (en) 2017-06-02 2020-05-19 Apple Inc. Multi-task recurrent neural network architecture for efficient morphology handling in neural language modeling
US10445429B2 (en) 2017-09-21 2019-10-15 Apple Inc. Natural language understanding using vocabularies with compressed serialized tries
US10755051B2 (en) 2017-09-29 2020-08-25 Apple Inc. Rule-based natural language processing
US10636424B2 (en) 2017-11-30 2020-04-28 Apple Inc. Multi-turn canned dialog
US10733982B2 (en) 2018-01-08 2020-08-04 Apple Inc. Multi-directional dialog
US10733375B2 (en) 2018-01-31 2020-08-04 Apple Inc. Knowledge-based framework for improving natural language understanding
US10789959B2 (en) 2018-03-02 2020-09-29 Apple Inc. Training speaker recognition models for digital assistants
US10592604B2 (en) 2018-03-12 2020-03-17 Apple Inc. Inverse text normalization for automatic speech recognition
US10818288B2 (en) 2018-03-26 2020-10-27 Apple Inc. Natural assistant interaction
US10909331B2 (en) 2018-03-30 2021-02-02 Apple Inc. Implicit identification of translation payload with neural machine translation
CN108401192B (en) * 2018-04-25 2022-02-22 腾讯科技(深圳)有限公司 Video stream processing method and device, computer equipment and storage medium
US10928918B2 (en) 2018-05-07 2021-02-23 Apple Inc. Raise to speak
US11145294B2 (en) 2018-05-07 2021-10-12 Apple Inc. Intelligent automated assistant for delivering content from user experiences
US10984780B2 (en) 2018-05-21 2021-04-20 Apple Inc. Global semantic word embeddings using bi-directional recurrent neural networks
US10892996B2 (en) 2018-06-01 2021-01-12 Apple Inc. Variable latency device coordination
DK201870355A1 (en) 2018-06-01 2019-12-16 Apple Inc. Virtual assistant operation in multi-device environments
DK180639B1 (en) 2018-06-01 2021-11-04 Apple Inc DISABILITY OF ATTENTION-ATTENTIVE VIRTUAL ASSISTANT
US11386266B2 (en) 2018-06-01 2022-07-12 Apple Inc. Text correction
DK179822B1 (en) 2018-06-01 2019-07-12 Apple Inc. Voice interaction at a primary device to access call functionality of a companion device
US10504518B1 (en) 2018-06-03 2019-12-10 Apple Inc. Accelerated task performance
US11100498B2 (en) 2018-06-03 2021-08-24 Apple Inc. User interfaces for transfer accounts
KR102447385B1 (en) 2018-06-03 2022-09-27 애플 인크. User Interfaces for Transfer Accounts
US11010561B2 (en) 2018-09-27 2021-05-18 Apple Inc. Sentiment prediction from textual data
US10839159B2 (en) 2018-09-28 2020-11-17 Apple Inc. Named entity normalization in a spoken dialog system
US11170166B2 (en) 2018-09-28 2021-11-09 Apple Inc. Neural typographical error modeling via generative adversarial networks
US11462215B2 (en) 2018-09-28 2022-10-04 Apple Inc. Multi-modal inputs for voice commands
US11475898B2 (en) 2018-10-26 2022-10-18 Apple Inc. Low-latency multi-speaker speech recognition
US11638059B2 (en) 2019-01-04 2023-04-25 Apple Inc. Content playback on multiple devices
US11348573B2 (en) 2019-03-18 2022-05-31 Apple Inc. Multimodality in digital assistant systems
US11328352B2 (en) 2019-03-24 2022-05-10 Apple Inc. User interfaces for managing an account
US20200320898A1 (en) * 2019-04-05 2020-10-08 Rally Reader, LLC Systems and Methods for Providing Reading Assistance Using Speech Recognition and Error Tracking Mechanisms
US11307752B2 (en) 2019-05-06 2022-04-19 Apple Inc. User configurable task triggers
US11475884B2 (en) 2019-05-06 2022-10-18 Apple Inc. Reducing digital assistant latency when a language is incorrectly determined
US11423908B2 (en) 2019-05-06 2022-08-23 Apple Inc. Interpreting spoken requests
DK201970509A1 (en) 2019-05-06 2021-01-15 Apple Inc Spoken notifications
US11140099B2 (en) 2019-05-21 2021-10-05 Apple Inc. Providing message response suggestions
US11289073B2 (en) 2019-05-31 2022-03-29 Apple Inc. Device text to speech
US11496600B2 (en) 2019-05-31 2022-11-08 Apple Inc. Remote execution of machine-learned models
DK201970510A1 (en) 2019-05-31 2021-02-11 Apple Inc Voice identification in digital assistant systems
DK180129B1 (en) 2019-05-31 2020-06-02 Apple Inc. User activity shortcut suggestions
US11360641B2 (en) 2019-06-01 2022-06-14 Apple Inc. Increasing the relevance of new available information
US11488406B2 (en) 2019-09-25 2022-11-01 Apple Inc. Text detection using global geometry estimators
US12118562B2 (en) 2020-05-29 2024-10-15 Apple Inc. Configuring an account for a second user identity
US11983702B2 (en) 2021-02-01 2024-05-14 Apple Inc. Displaying a representation of a card with a layered structure
US11921992B2 (en) 2021-05-14 2024-03-05 Apple Inc. User interfaces related to time
US11784956B2 (en) 2021-09-20 2023-10-10 Apple Inc. Requests to add assets to an asset account
US12130965B2 (en) * 2022-07-07 2024-10-29 Plume Design, Inc. Ring enabling its wearer to enter control commands

Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4195220A (en) * 1977-11-21 1980-03-25 Bristol Stanley M Portable elapsed time recorder
US4724285A (en) * 1985-11-07 1988-02-09 Digitext, Inc. Stenographic translation system
US4984274A (en) * 1988-07-07 1991-01-08 Casio Computer Co., Ltd. Speech recognition apparatus with means for preventing errors due to delay in speech recognition
US5369704A (en) * 1993-03-24 1994-11-29 Engate Incorporated Down-line transcription system for manipulating real-time testimony
US5606497A (en) * 1994-03-30 1997-02-25 Cramer; Milton L. Method and apparatus for recording billable time and services
US5721827A (en) * 1996-10-02 1998-02-24 James Logan System for electrically distributing personalized information
US5855000A (en) * 1995-09-08 1998-12-29 Carnegie Mellon University Method and apparatus for correcting and repairing machine-transcribed input using independent or cross-modal secondary input
US5991742A (en) * 1996-05-20 1999-11-23 Tran; Bao Q. Time and expense logging system
US6304851B1 (en) * 1998-03-13 2001-10-16 The Coca-Cola Company Mobile data collection systems, methods and computer program products
US20010049470A1 (en) * 2000-01-19 2001-12-06 Mault James R. Diet and activity monitoring device
US6539359B1 (en) * 1998-10-02 2003-03-25 Motorola, Inc. Markup language for interactive services and methods thereof
US6675142B2 (en) * 1999-06-30 2004-01-06 International Business Machines Corporation Method and apparatus for improving speech recognition accuracy
US7127395B1 (en) * 2001-01-22 2006-10-24 At&T Corp. Method and system for predicting understanding errors in a task classification system
US7171365B2 (en) * 2001-02-16 2007-01-30 International Business Machines Corporation Tracking time using portable recorders and speech recognition

Patent Citations (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4195220A (en) * 1977-11-21 1980-03-25 Bristol Stanley M Portable elapsed time recorder
US4724285A (en) * 1985-11-07 1988-02-09 Digitext, Inc. Stenographic translation system
US4984274A (en) * 1988-07-07 1991-01-08 Casio Computer Co., Ltd. Speech recognition apparatus with means for preventing errors due to delay in speech recognition
US5369704A (en) * 1993-03-24 1994-11-29 Engate Incorporated Down-line transcription system for manipulating real-time testimony
US5926787A (en) * 1993-03-24 1999-07-20 Engate Incorporated Computer-aided transcription system using pronounceable substitute text with a common cross-reference library
US5606497A (en) * 1994-03-30 1997-02-25 Cramer; Milton L. Method and apparatus for recording billable time and services
US5855000A (en) * 1995-09-08 1998-12-29 Carnegie Mellon University Method and apparatus for correcting and repairing machine-transcribed input using independent or cross-modal secondary input
US5991742A (en) * 1996-05-20 1999-11-23 Tran; Bao Q. Time and expense logging system
US5721827A (en) * 1996-10-02 1998-02-24 James Logan System for electrically distributing personalized information
US6304851B1 (en) * 1998-03-13 2001-10-16 The Coca-Cola Company Mobile data collection systems, methods and computer program products
US6539359B1 (en) * 1998-10-02 2003-03-25 Motorola, Inc. Markup language for interactive services and methods thereof
US6675142B2 (en) * 1999-06-30 2004-01-06 International Business Machines Corporation Method and apparatus for improving speech recognition accuracy
US20010049470A1 (en) * 2000-01-19 2001-12-06 Mault James R. Diet and activity monitoring device
US7127395B1 (en) * 2001-01-22 2006-10-24 At&T Corp. Method and system for predicting understanding errors in a task classification system
US7171365B2 (en) * 2001-02-16 2007-01-30 International Business Machines Corporation Tracking time using portable recorders and speech recognition

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8719121B1 (en) * 2008-01-15 2014-05-06 David Leason System and method for automated construction of time records based on electronic messages
US20100010814A1 (en) * 2008-07-08 2010-01-14 International Business Machines Corporation Enhancing media playback with speech recognition
US8478592B2 (en) * 2008-07-08 2013-07-02 Nuance Communications, Inc. Enhancing media playback with speech recognition
US10475464B2 (en) * 2012-07-03 2019-11-12 Samsung Electronics Co., Ltd Method and apparatus for connecting service between user devices using voice
US20180034961A1 (en) * 2014-02-28 2018-02-01 Ultratec, Inc. Semiautomated Relay Method and Apparatus
US12137183B2 (en) 2014-02-28 2024-11-05 Ultratec, Inc. Semiautomated relay method and apparatus
US12136425B2 (en) 2014-02-28 2024-11-05 Ultratec, Inc. Semiautomated relay method and apparatus
US12136426B2 (en) 2014-02-28 2024-11-05 Ultratec, Inc. Semiautomated relay method and apparatus
US12035070B2 (en) 2020-02-21 2024-07-09 Ultratec, Inc. Caption modification and augmentation systems and methods for use by hearing assisted user

Also Published As

Publication number Publication date
US7171365B2 (en) 2007-01-30
US20020116185A1 (en) 2002-08-22
US20080288251A1 (en) 2008-11-20
US7664638B2 (en) 2010-02-16

Similar Documents

Publication Publication Date Title
US7664638B2 (en) Tracking time using portable recorders and speech recognition
US8335689B2 (en) Method and system for efficient management of speech transcribers
US7716054B2 (en) Activity-ware for non-textual objects
US5526407A (en) Method and apparatus for managing information
US7047192B2 (en) Simultaneous multi-user real-time speech recognition system
US6839669B1 (en) Performing actions identified in recognized speech
US7236932B1 (en) Method of and apparatus for improving productivity of human reviewers of automatically transcribed documents generated by media conversion systems
US8050923B2 (en) Automated utterance search
US7797331B2 (en) Method and device for organizing user provided information with meta-information
US20050102146A1 (en) Method and apparatus for voice dictation and document production
US20050187839A1 (en) Realtime billable timekeeper method, system and apparatus
WO2000031724A1 (en) System and method for automating transcription services
EP1917586A2 (en) Recorded customer interactions and training system, method and computer program product
JPH06110755A (en) System for indexing data set
Whittaker et al. Design and evaluation of systems to support interaction capture and retrieval
JP2004534326A (en) Method of providing settlement information and method and apparatus for dictation
CN117912466A (en) Virtual duty method for intelligent control of auxiliary energy service domain based on artificial intelligence
CN115034186A (en) Report generation method and system based on data processing
US6728676B1 (en) Using speech recognition to improve efficiency of an inventory task
US20230188643A1 (en) Ai-based real-time natural language processing system and method thereof
US9047872B1 (en) Automatic speech recognition tuning management
JP3234083B2 (en) Search device
US20070078806A1 (en) Method and apparatus for evaluating the accuracy of transcribed documents and other documents
JP3469561B2 (en) Search device
US20220277733A1 (en) Real-time communication and collaboration system and method of monitoring objectives to be achieved by a plurality of users collaborating on a real-time communication and collaboration platform

Legal Events

Date Code Title Description
AS Assignment

Owner name: NUANCE COMMUNICATIONS, INC., MASSACHUSETTS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:INTERNATIONAL BUSINESS MACHINES CORPORATION;REEL/FRAME:022330/0088

Effective date: 20081231

Owner name: NUANCE COMMUNICATIONS, INC.,MASSACHUSETTS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:INTERNATIONAL BUSINESS MACHINES CORPORATION;REEL/FRAME:022330/0088

Effective date: 20081231

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO PAY ISSUE FEE