US20020027976A1 - System for recording, annotating and indexing audio data - Google Patents

System for recording, annotating and indexing audio data Download PDF

Info

Publication number
US20020027976A1
US20020027976A1 US09/058,283 US5828398A US2002027976A1 US 20020027976 A1 US20020027976 A1 US 20020027976A1 US 5828398 A US5828398 A US 5828398A US 2002027976 A1 US2002027976 A1 US 2002027976A1
Authority
US
United States
Prior art keywords
audio data
property
user
audio
keyword
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US09/058,283
Other versions
US6404856B1 (en
Inventor
Lynn D. Wilcox
Patrick R. Chiu
Jacek Gwizdka
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Majandro LLC
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US09/058,283 priority Critical patent/US6404856B1/en
Assigned to XEROX CORPORATION, FUJI XEROX CO., LTD. reassignment XEROX CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: GWIZDKA, JACEK, CHIU, PATRICK R., WILCOX, LYNN D.
Priority to JP11089695A priority patent/JPH11313173A/en
Priority to EP99302805A priority patent/EP0949621A3/en
Publication of US20020027976A1 publication Critical patent/US20020027976A1/en
Application granted granted Critical
Publication of US6404856B1 publication Critical patent/US6404856B1/en
Assigned to BANK ONE, NA, AS ADMINISTRATIVE AGENT reassignment BANK ONE, NA, AS ADMINISTRATIVE AGENT SECURITY AGREEMENT Assignors: XEROX CORPORATION
Assigned to JPMORGAN CHASE BANK, AS COLLATERAL AGENT reassignment JPMORGAN CHASE BANK, AS COLLATERAL AGENT SECURITY AGREEMENT Assignors: XEROX CORPORATION
Assigned to XEROX CORPORATION reassignment XEROX CORPORATION RELEASE BY SECURED PARTY (SEE DOCUMENT FOR DETAILS). Assignors: BANK ONE, NA
Assigned to XEROX CORPORATION reassignment XEROX CORPORATION RELEASE BY SECURED PARTY (SEE DOCUMENT FOR DETAILS). Assignors: JPMORGAN CHASE BANK, N.A.
Assigned to XEROX CORPORATION reassignment XEROX CORPORATION CORRECTIVE ASSIGNMENT TO CORRECT THE ASSIGNEE'S NAME PREVIOUSLY RECORDED AT REEL: 009411 FRAME: 0319. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT. Assignors: GWIZDKA, JACEK, CHIU, PATRICK R, WILCOX, LYNN D
Anticipated expiration legal-status Critical
Assigned to MAJANDRO LLC reassignment MAJANDRO LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: XEROX CORPORATION
Assigned to XEROX CORPORATION reassignment XEROX CORPORATION RELEASE BY SECURED PARTY (SEE DOCUMENT FOR DETAILS). Assignors: JPMORGAN CHASE BANK, N.A. AS SUCCESSOR-IN-INTEREST ADMINISTRATIVE AGENT AND COLLATERAL AGENT TO BANK ONE, N.A.
Assigned to XEROX CORPORATION reassignment XEROX CORPORATION RELEASE BY SECURED PARTY (SEE DOCUMENT FOR DETAILS). Assignors: JPMORGAN CHASE BANK, N.A. AS SUCCESSOR-IN-INTEREST ADMINISTRATIVE AGENT AND COLLATERAL AGENT TO JPMORGAN CHASE BANK
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/64Automatic arrangements for answering calls; Automatic arrangements for recording messages for absent subscribers; Arrangements for recording conversations
    • H04M1/65Recording arrangements for recording a message from the calling party
    • H04M1/6505Recording arrangements for recording a message from the calling party storing speech in digital form
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/102Programmed access in sequence to addressed parts of tracks of operating record carriers
    • G11B27/105Programmed access in sequence to addressed parts of tracks of operating record carriers of operating discs
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/19Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
    • G11B27/28Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/34Indicating arrangements 
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/57Arrangements for indicating or recording the number of the calling subscriber at the called subscriber's set
    • H04M1/575Means for retrieving and displaying personal data about calling party
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/64Automatic arrangements for answering calls; Automatic arrangements for recording messages for absent subscribers; Arrangements for recording conversations
    • H04M1/65Recording arrangements for recording a message from the calling party
    • H04M1/656Recording arrangements for recording a message from the calling party for recording conversations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M3/00Automatic or semi-automatic exchanges
    • H04M3/42Systems providing special services or facilities to subscribers
    • H04M3/42221Conversation recording systems
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B2220/00Record carriers by type
    • G11B2220/20Disc-shaped record carriers
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B2220/00Record carriers by type
    • G11B2220/20Disc-shaped record carriers
    • G11B2220/21Disc-shaped record carriers characterised in that the disc is of read-only, rewritable, or recordable type
    • G11B2220/213Read-only discs
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B2220/00Record carriers by type
    • G11B2220/20Disc-shaped record carriers
    • G11B2220/25Disc-shaped record carriers characterised in that the disc is based on a specific recording technology
    • G11B2220/2508Magnetic discs
    • G11B2220/2512Floppy disks
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B2220/00Record carriers by type
    • G11B2220/20Disc-shaped record carriers
    • G11B2220/25Disc-shaped record carriers characterised in that the disc is based on a specific recording technology
    • G11B2220/2537Optical discs
    • G11B2220/2545CDs
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B2220/00Record carriers by type
    • G11B2220/60Solid state media
    • G11B2220/61Solid state media wherein solid state memory is used for storing A/V content
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2201/00Electronic components, circuits, software, systems or apparatus used in telephone systems
    • H04M2201/38Displays
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M3/00Automatic or semi-automatic exchanges
    • H04M3/42Systems providing special services or facilities to subscribers
    • H04M3/42025Calling or Called party identification service
    • H04M3/42034Calling party identification service
    • H04M3/42042Notifying the called party of information on the calling party
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M3/00Automatic or semi-automatic exchanges
    • H04M3/42Systems providing special services or facilities to subscribers
    • H04M3/42025Calling or Called party identification service
    • H04M3/42034Calling party identification service
    • H04M3/42059Making use of the calling party identifier
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M3/00Automatic or semi-automatic exchanges
    • H04M3/42Systems providing special services or facilities to subscribers
    • H04M3/436Arrangements for screening incoming calls, i.e. evaluating the characteristics of a call before deciding whether to answer it
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M3/00Automatic or semi-automatic exchanges
    • H04M3/42Systems providing special services or facilities to subscribers
    • H04M3/50Centralised arrangements for answering calls; Centralised arrangements for recording messages for absent or busy subscribers ; Centralised arrangements for recording messages
    • H04M3/53Centralised arrangements for recording incoming messages, i.e. mailbox systems
    • H04M3/533Voice mail systems

Definitions

  • This invention is directed to a system for recording, annotating and indexing audio data.
  • the invention is directed to an audio system that improves the way users record, review and retrieve audio data.
  • the '311 application discloses an electronic notebook that improves the way users record, review, retrieve and reuse their personal notes.
  • the electronic notebook includes an interface that is similar to a standard paper notebook. That is, the user enters freeform hand-written notes onto a screen having the appearance of a page into a digital notebook.
  • the electronic notebook also has features that augment standard note-taking capabilities.
  • the electronic notebook allows a property, such as “action item” or “name” to be associated with the note marks so that notes with these properties may be more easily retrieved.
  • the electronic notebook allows the user to assign text keywords to a page of notes. Finally, the electronic notebook records the audio data and synchronizes the audio data with the note marks. However, only those portions of the audio data that is highlighted by the user are permanently stored.
  • These features of the electronic notebook provide the ability to have multiple views on a notebook. These multiple view arc analogous to views available in a database system. For example, requesting notes from last week with the property “action item” and keywords “mobile” and “computing” causes the notebook to display only those note marks from a previous week which were marked as action items and which were on the topic of mobile computing. Once a notebook's views are defined they may be reused. This provides the user with dynamic organization of the changing content of the notebook.
  • the note marks are linked to the audio data, so that the user can play the audio data that was recorded at the time a particular note was written. Audio data is displayed with an audio timeline at the top of the note page. Important audio data can be highlighted. Audio data highlights are shown as colored bands in the audio display that can be selected for playback.
  • Note marks can be indexed using ink properties. Ink properties are data types that are associated with particular types of ink marks. The ink properties arc indicated to the user by ink color and/or mark size.
  • pages or note marks can be indexed by keywords. Note marks can be retrieved by queries using properties and keywords. Property queries retrieve only those note marks having the specified property, while keyword queries retrieve an entire page of note marks. The electronic notebook disclosed in the '311 application, however, does allow access to the audio data without referring to the associated note marks.
  • the Apple Newton® places phone calls, and provides a note page to take notes on the conversation. However, the conversation is not recorded.
  • the Apple Newton® also records messages and sounds but does not associate properties, text or ink to the recording.
  • This invention provides a method and apparatus that uses keywords and properties to index audio data.
  • This invention further provides a method and apparatus that separates the sides of a conversation in the audio data to improve access.
  • This invention additionally provides a method and apparatus for recording, indexing and retrieving voice mail messages.
  • the method and apparatus of this invention applies properties to the audio data and generates indexes to the audio data. This indexing allows retrieval of specific audio data and portions of audio data.
  • This invention allows users to take notes using a computer while recording the audio data of a telephone call.
  • the two sides of the conversation are separated and shown in an audio display for easy access to specific comments using the associated audio data attributes, such as date and time, properties, keywords, and identity of the caller. These associated attributes are used to index the audio data for subsequent retrieval. Properties apply to the audio data as well as to note marks to provide access to specific remarks in the audio data.
  • This invention allows a user to play specific comments from previous calls during a current conversation. This is useful when the tone of voice used in a call conveys information that is not available in the associated note marks or text.
  • Properties for the audio data can be set without selecting or writing note marks by, for example, pressing a button for the desired property.
  • the property of the audio data is displayed to the user by coloring the band for the current speaker turn in the audio bar.
  • the method and apparatus of this invention assume that the extent of each speaker's turn corresponds to the extent of the selected audio data property.
  • the method and apparatus of this invention route calls based upon a mode. Depending on the mode, the method and apparatus of this invention send calls to an answering machine, accept a call for conversation, or tentatively accept a call based on the identity of the caller.
  • FIG. 1 is a block diagram of one embodiment of an audio data system according to this invention.
  • FIG. 2 shows one embodiment of the audio data system incorporated into a telephone set in accordance with this invention
  • FIG. 3 illustrates a user interface of one embodiment of the audio data system according to this invention
  • FIG. 4 illustrates a sample retrieval query window of the audio data system according to this invention.
  • FIG. 5 illustrates a sample table of contents of retrieved audio data files according to this invention.
  • FIG. 1 is a block diagram of one embodiment of an audio data system 10 according to this invention.
  • the audio data system 10 has a processor 12 communicating with a storage device 14 , a memory 16 and an input/output interface 18 .
  • a user interacts with and controls the audio data system 10 through any number of conventional input/output devices, such as a pen 20 , a keyboard 22 , a mouse 24 , a microphone 26 , an audio data input device 28 , a display 30 and a speaker 32 .
  • the input/output devices 20 - 32 communicate with the input/output interface 18 that, in turn, communicates with the processor 12 .
  • the audio data system 10 is preferably implemented on a programmed general purpose computer. However, the system 10 can also be implemented using a specific purpose computer, a programmed microprocessor or microcontroller and any necessary peripheral integrated circuit elements, an ASIC or other integrated circuit, a hardwired electronic or logic circuit such as a discrete element circuit, a digital signal processor, a programmable logic device such as a PLD, PLA, FPGA, or PAL, or the like.
  • An alternate portion of the memory 16 is preferably implemented using static or dynamic RAM, and stores executing application programs and data.
  • the alterable portion of the memory 16 can also be implemented using a floppy disk and drive disk, a writable optical disk and disk drive, a hard drive, flash memory or the like.
  • a fixed portion of the memory 16 is preferably implemented using a ROM and stores basic control programs for the audio data system.
  • the fixed portion of the memory can also be implemented using a PROM, an EPROM, an EEPROM, or a CD-ROM and disk drive or the like.
  • the storage device 14 is preferably implemented using a hard drive, and provides long-term storage for application programs and data. However, the storage device 14 can also be implemented using a floppy disk and disk drive, a writable optical disk and disk drive, flash memory, static or dynamic RAM, or the like.
  • the telephone system connector 28 may be any type of connector that is capable of carrying audio data.
  • the telephone system connector 28 can be a connection to a publicly or privately switched telephone network, a wired or wireless link to a data network.
  • the data network can be a local area network, a wide area network, an intranet, the Internet or any other distributed processing and storage network.
  • FIG. 2 shows one embodiment of the audio data system 10 of this invention that has been incorporated into a telephone 34 .
  • the audio data system 10 in FIG. 2 has an audio data input device 28 that is adapted to receive audio data from a telephone system and a handset and/or a microphone 26 that have been built into the telephone for hands free operation.
  • This audio data system 10 also includes a keypad 36 and a pen-based input display 20 which are used as a user interface to control the operation of the audio data system 10 .
  • the audio data system 10 of this invention allows a user to indicate the type of information contained in audio data by attaching properties to the audio data.
  • Properties indicate information types that are associated with the selected audio data. Properties are associated with portions of the audio data, not the entire audio data file, so that individual portions of the audio data can be easily identified. Audio data portions having a specific property from a number of audio data files can be grouped and listed together. Properties are useful for retrieval because they are a low overhead categorization that can be used to distinguish between audio data portions.
  • the audio data system 10 of this invention records audio data and links the audio data to handwritten notes.
  • the sides of the conversation in a telephone call are separated using a speaker segmentation algorithm.
  • the resulting speaker segmentation is shown in the audio timeline bar 40 at the top of the note page 42 in FIG. 3.
  • a local speaker is shown by the dark bands in the top bar 44 and another speaker is shown in the lower bar 46 .
  • Speaker segmentation and display can be extended to any number of sides of a conversation. If the audio data system 10 is connected to a conventional telephone system an elaborate speaker segmentation algorithm may be avoided by relying on the fact that each party will be using separate channels. All incoming audio can be attributed to a remote speaker and all outgoing audio can be attributed to a local speaker.
  • Audio data portions having different properties are distinguished from each other by the color of the timeline 44 . For example, audio data portions that do not have any assigned properties are shown in black, while audio data portions having “action item” as an assigned property are shown in green.
  • the audio data system 10 allows properties to be applied to the audio data as well as to ink. Specific audio data can be selected and a property applied to that audio data.
  • the property is indicated in the audio bar 40 by changing the color of the audio bar 40 to a color associated with that property. For example, the “to do” property may be indicated to the user in blue ink for the corresponding note marks and as a blue band in the audio bar 40 for the corresponding audio data.
  • Properties can be assigned to the audio data portions either before or after the audio data is recorded. If the user knows beforehand that the audio data about to be recorded has a certain property, the user can select that property from a menu, assign it to the audio data and begin recording. Properties may be assigned after recording by selecting the audio and assigning a property to the selected audio. For example, toggling one of the property buttons 50 , a property corresponding to the property button is applied to the current turn in the conversation, indicating that this audio data portion has the given property.
  • a property button 52 which does not assign a property
  • a topic property button 54 a name property button 56 , a quote property button 58 , a phone property button 60 , a to do property button 62 , a Uniform Resource Locator (URL) property button 64 , an organization property button 66 and a user defined property button.
  • Properties can be assigned to audio data and note marks that are currently being received or which have been selected by toggling the property button 50 which corresponds to the desired property.
  • the audio data system 10 may also indicate an audio data property, even though there are no associated note marks with that property.
  • the identity of a caller is used to index the audio data.
  • This index can be obtained in several ways.
  • the caller can be identified using a database linking a phone number or a caller id to the caller. If the caller is not calling from his own phone, the audio data system 10 of this invention uses a database of audio data files of known callers, and matches the voice from the incoming caller to an audio data file of a known caller. Any conventional speaker identification system may be used.
  • a speaker identification system is disclosed in U.S. Pat. No. 5,655,058, herein incorporated by reference in its entirety.
  • the user can also enter the caller's identity through any other conventional input method.
  • the audio data can also be indexed automatically according to date and time.
  • FIG. 3 shows a call initiation dialog window 70 on top of user interface 42 .
  • a user can access the call initiation dialog window 70 by selecting the call initiation button 72 on the user interface 42 .
  • the call initiation dialog window 70 includes a keypad 74 that enables the user to manually enter a telephone number.
  • the call initiation dialog window 70 also includes a phonebook button 76 which provides access to a phonebook listing of telephone numbers (not shown).
  • the call initiation dialog window 70 also has a dropdown text entry box 78 which enables a user to place a cursor in the text entry box 78 and enter text and/or numbers using a separate data input device such as a keyboard.
  • the call initiation dialog window 70 has a cancel button 80 which enables a user to cancel a call initiation and also an “ok” button 82 which enables a user to initiate a call once the telephone number information has been entered.
  • This invention also provides a system for call screening and/or routing based on speaker identity.
  • a screen mode (not shown) of the audio data system 10 can be set so that all calls are automatically answered, all calls are automatically recorded without interaction (as in voice mail), or calls are routed in accordance with the identity of the caller.
  • the audio data system 10 answers the call and sends a message to the caller to state his name and the purpose of the call. If the caller's identity matches a name on an “accept call” list, the user is notified of the call. Otherwise, the call is recorded as voice mail.
  • the audio data system 10 of this invention allows keywords to be associated with audio data. This allows keyword-based retrieval of audio data.
  • keywords describe the actual content of the audio data. Because keywords reflect more general information, they are assigned to the entire audio data file and not to particular portions of audio data. Keywords are displayed in the margins of a digital page (not shown) and can be entered in several ways. For example, keywords can be selected from a list of previously entered keywords. Alternately, keywords can be typed in using a pop-up keyboard (not shown). Additionally, keywords can be hand-written in a keyword entry box (not shown) and recognized using a standard handwriting recognition system.
  • note marks can be selected and sent to a handwriting recognizer, and the resulting text entered as a keyword for the audio data associated with the note marks. If the user does not want to immediately interact with the handwriting recognition system, the user can associate the “keyword” property with digital ink to cause delayed recognition.
  • the audio data can be retrieved using attributes associated with the audio data such as caller, time, and topic, using properties and keywords.
  • the user interface 42 also has query buttons 112 that enable a user to access a context-based query (not shown) by selecting the context-based query button 114 or access a query window 84 as shown in FIG. 4 using the query button 116 .
  • the query window 84 includes property check boxes 86 that enable a user to search based upon a designated property.
  • the query window 84 also has a keyword entry box 88 that permits a user to search for audio data and notemarks having a specified keyword. Retrieval by keyword and name returns a list of audio data files with the identified caller and having the desired keyword.
  • Retrieval by property is more specific, since it provides the audio data portions of calls and digital ink groups having the specified property. This is convenient for locating a particular comment in an audio data file. For example, if a comment is marked with the property “important”, that audio data segment containing the comment can be retrieved using a query with the “important” property designated. Segments of audio data files can also be played to both parties during a conversation. Thus, comments from previous conversations can be shared.
  • Results from a search are shown in FIG. 5 in a search result window 94 .
  • the search result window lists the attributes of the retrieved data files in columns. The columns indicate the creation date 96 , the date of last modification 98 , the type of property as indicated by icon 100 and the keywords 102 .
  • the series of icons describe properties associated with the audio data files. For example, the telephone icon represents the property “phone number” and the check box represents the “action item” property.
  • the final column displays keywords associated with the audio data files.
  • the audio data system of this invention provides the user with multiple views of audio data stored in a notebook.
  • the views provided by this invention are analogous to the views in a database system. If a query specifies a particular property, only those audio data portions having that property will be displayed. This view of the audio data is formed by newly generated audio data files, each having an original date and containing only the audio data portions with the specified property.
  • the user can also create a view of a notebook by forming a query to request audio data files having specific keywords.
  • the user can define the keywords in the query by browsing an alphabetized list of keywords that have been attached to pages, or by typing or writing the desired keywords.
  • views are indicated by tabs 48 at the bottom of the notebook page.
  • a name is associated with each tab 48 for easy reference.
  • the user can return to the previous view by selecting the corresponding tab 48 .
  • the user can decide if the view should be static or dynamic. When the view is static, the contents of the view remains the same. When the view is dynamic, new items are added to the view as they are recorded by the notebook, provided that these items satisfy the terms of the query for this view.
  • a user can navigate through a notebook using the navigation buttons 104 in the user interface window 42 .
  • Navigation buttons 106 and 108 allow a user to page forward and backward through pages in a notebook.
  • Navigation Button 110 is a new page button that starts data entry on a fresh notebook page.
  • the user interface 42 is also provided with mode select buttons 118 which enable a user to select either a page mode using a page mode select button 120 or a table of contents mode using the table of contents mode select button 122 .
  • the user interface generally includes the features specified and described in co-pending U.S. patent application Ser. No. 08/821,311.

Abstract

An audio data system assigns properties to audio data and keywords to audio data files. The properties indicate the type of audio data and the keywords indicate the general content of an audio data file. The system provides an updatable database of audio data that is capable of being searched based on the properties and the keywords.

Description

    BACKGROUND OF THE INVENTION
  • 1. Field of Invention [0001]
  • This invention is directed to a system for recording, annotating and indexing audio data. In particular, the invention is directed to an audio system that improves the way users record, review and retrieve audio data. [0002]
  • 2. Description of Related Art [0003]
  • Many people take notes on their telephone conversations. Interviews with lawyers, managers, administrative assistants, salesmen and executives have shown that certain note-taking practices are used during phone calls. Some people record the date and time of the call, as well as the identity of the caller, some people keep a separate notebook by the phone for notes on their calls and others keep a special notebook for notes taken on voice mail messages. People take notes on information obtained during their calls, generate a list of “to do” items, and record other details such as sales orders and directions. If an administrative assistant answers a call for his boss and finds the message too complex, the call is transferred to the voice mail recording system. [0004]
  • Recording a call increases the accuracy of the information that is obtained from the call. However, while voice mail systems can record phone conversations, they do not index the audio data for later access. [0005]
  • U.S. patent application Ser. No. 08/821,311, filed Mar. 20, 1997, which is incorporated herein in its entirety, discloses a system that allows users to simultaneously record audio data while taking hand-written notes. The '311 application discloses an electronic notebook that improves the way users record, review, retrieve and reuse their personal notes. The electronic notebook includes an interface that is similar to a standard paper notebook. That is, the user enters freeform hand-written notes onto a screen having the appearance of a page into a digital notebook. The electronic notebook also has features that augment standard note-taking capabilities. The electronic notebook allows a property, such as “action item” or “name” to be associated with the note marks so that notes with these properties may be more easily retrieved. In order to further aid indexing, the electronic notebook allows the user to assign text keywords to a page of notes. Finally, the electronic notebook records the audio data and synchronizes the audio data with the note marks. However, only those portions of the audio data that is highlighted by the user are permanently stored. [0006]
  • These features of the electronic notebook provide the ability to have multiple views on a notebook. These multiple view arc analogous to views available in a database system. For example, requesting notes from last week with the property “action item” and keywords “mobile” and “computing” causes the notebook to display only those note marks from a previous week which were marked as action items and which were on the topic of mobile computing. Once a notebook's views are defined they may be reused. This provides the user with dynamic organization of the changing content of the notebook. [0007]
  • The note marks are linked to the audio data, so that the user can play the audio data that was recorded at the time a particular note was written. Audio data is displayed with an audio timeline at the top of the note page. Important audio data can be highlighted. Audio data highlights are shown as colored bands in the audio display that can be selected for playback. Note marks can be indexed using ink properties. Ink properties are data types that are associated with particular types of ink marks. The ink properties arc indicated to the user by ink color and/or mark size. In addition, pages or note marks can be indexed by keywords. Note marks can be retrieved by queries using properties and keywords. Property queries retrieve only those note marks having the specified property, while keyword queries retrieve an entire page of note marks. The electronic notebook disclosed in the '311 application, however, does allow access to the audio data without referring to the associated note marks. [0008]
  • “Capturing, Structuring and Representing Ubiquitous Audio”, by D. Hindus et al., ACM Trans. Information Systems, Vol. II, No. 4, October 1993, pp. 376-400, discloses a system that records telephone calls while displaying the turn-taking structure of the conversation. During the telephone call, the user selects parts of the conversation to save. After the call, the user can add text tags to the stored parts of the conversation. However, this system does not permit the user to enter pen-based input. This system only permits typed input and then only after the telephone call. [0009]
  • The Apple Newton® places phone calls, and provides a note page to take notes on the conversation. However, the conversation is not recorded. The Apple Newton® also records messages and sounds but does not associate properties, text or ink to the recording. [0010]
  • SUMMARY OF THE INVENTION
  • This invention provides a method and apparatus that uses keywords and properties to index audio data. [0011]
  • This invention further provides a method and apparatus that separates the sides of a conversation in the audio data to improve access. [0012]
  • This invention additionally provides a method and apparatus for recording, indexing and retrieving voice mail messages. [0013]
  • In particular, the method and apparatus of this invention, applies properties to the audio data and generates indexes to the audio data. This indexing allows retrieval of specific audio data and portions of audio data. [0014]
  • This invention allows users to take notes using a computer while recording the audio data of a telephone call. The two sides of the conversation are separated and shown in an audio display for easy access to specific comments using the associated audio data attributes, such as date and time, properties, keywords, and identity of the caller. These associated attributes are used to index the audio data for subsequent retrieval. Properties apply to the audio data as well as to note marks to provide access to specific remarks in the audio data. [0015]
  • This invention allows a user to play specific comments from previous calls during a current conversation. This is useful when the tone of voice used in a call conveys information that is not available in the associated note marks or text. [0016]
  • Properties for the audio data can be set without selecting or writing note marks by, for example, pressing a button for the desired property. The property of the audio data is displayed to the user by coloring the band for the current speaker turn in the audio bar. The method and apparatus of this invention assume that the extent of each speaker's turn corresponds to the extent of the selected audio data property. [0017]
  • The method and apparatus of this invention route calls based upon a mode. Depending on the mode, the method and apparatus of this invention send calls to an answering machine, accept a call for conversation, or tentatively accept a call based on the identity of the caller. [0018]
  • These and other features and advantages of this invention are described in or are apparent from the following detailed description of the preferred embodiments. [0019]
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The preferred embodiments of this invention will be described in detail, with reference to the following figures, wherein: [0020]
  • FIG. 1 is a block diagram of one embodiment of an audio data system according to this invention; [0021]
  • FIG. 2 shows one embodiment of the audio data system incorporated into a telephone set in accordance with this invention; [0022]
  • FIG. 3 illustrates a user interface of one embodiment of the audio data system according to this invention; [0023]
  • FIG. 4 illustrates a sample retrieval query window of the audio data system according to this invention; and [0024]
  • FIG. 5 illustrates a sample table of contents of retrieved audio data files according to this invention.[0025]
  • These and other features and advantages of this invention are described in or are apparent from the following detailed description of the preferred embodiments. [0026]
  • DETAILED DESCRIPTION OF PREFERRED EMBODIMENTS
  • FIG. 1 is a block diagram of one embodiment of an [0027] audio data system 10 according to this invention. The audio data system 10 has a processor 12 communicating with a storage device 14, a memory 16 and an input/output interface 18. A user interacts with and controls the audio data system 10 through any number of conventional input/output devices, such as a pen 20, a keyboard 22, a mouse 24, a microphone 26, an audio data input device 28, a display 30 and a speaker 32. The input/output devices 20-32 communicate with the input/output interface 18 that, in turn, communicates with the processor 12.
  • The [0028] audio data system 10 is preferably implemented on a programmed general purpose computer. However, the system 10 can also be implemented using a specific purpose computer, a programmed microprocessor or microcontroller and any necessary peripheral integrated circuit elements, an ASIC or other integrated circuit, a hardwired electronic or logic circuit such as a discrete element circuit, a digital signal processor, a programmable logic device such as a PLD, PLA, FPGA, or PAL, or the like.
  • An alternate portion of the [0029] memory 16 is preferably implemented using static or dynamic RAM, and stores executing application programs and data. However, the alterable portion of the memory 16 can also be implemented using a floppy disk and drive disk, a writable optical disk and disk drive, a hard drive, flash memory or the like. A fixed portion of the memory 16 is preferably implemented using a ROM and stores basic control programs for the audio data system. However, the fixed portion of the memory can also be implemented using a PROM, an EPROM, an EEPROM, or a CD-ROM and disk drive or the like.
  • The [0030] storage device 14 is preferably implemented using a hard drive, and provides long-term storage for application programs and data. However, the storage device 14 can also be implemented using a floppy disk and disk drive, a writable optical disk and disk drive, flash memory, static or dynamic RAM, or the like.
  • The [0031] telephone system connector 28 may be any type of connector that is capable of carrying audio data. For example, the telephone system connector 28 can be a connection to a publicly or privately switched telephone network, a wired or wireless link to a data network. The data network can be a local area network, a wide area network, an intranet, the Internet or any other distributed processing and storage network.
  • FIG. 2 shows one embodiment of the [0032] audio data system 10 of this invention that has been incorporated into a telephone34. The audio data system 10 in FIG. 2 has an audio data input device 28 that is adapted to receive audio data from a telephone system and a handset and/or a microphone 26 that have been built into the telephone for hands free operation. This audio data system 10 also includes a keypad 36 and a pen-based input display 20 which are used as a user interface to control the operation of the audio data system 10.
  • The [0033] audio data system 10 of this invention allows a user to indicate the type of information contained in audio data by attaching properties to the audio data. Properties indicate information types that are associated with the selected audio data. Properties are associated with portions of the audio data, not the entire audio data file, so that individual portions of the audio data can be easily identified. Audio data portions having a specific property from a number of audio data files can be grouped and listed together. Properties are useful for retrieval because they are a low overhead categorization that can be used to distinguish between audio data portions.
  • The [0034] audio data system 10 of this invention records audio data and links the audio data to handwritten notes. The sides of the conversation in a telephone call are separated using a speaker segmentation algorithm. The resulting speaker segmentation is shown in the audio timeline bar 40 at the top of the note page 42 in FIG. 3. A local speaker is shown by the dark bands in the top bar 44 and another speaker is shown in the lower bar 46. Speaker segmentation and display can be extended to any number of sides of a conversation. If the audio data system 10 is connected to a conventional telephone system an elaborate speaker segmentation algorithm may be avoided by relying on the fact that each party will be using separate channels. All incoming audio can be attributed to a remote speaker and all outgoing audio can be attributed to a local speaker.
  • Audio data portions having different properties are distinguished from each other by the color of the [0035] timeline 44. For example, audio data portions that do not have any assigned properties are shown in black, while audio data portions having “action item” as an assigned property are shown in green.
  • The [0036] audio data system 10 allows properties to be applied to the audio data as well as to ink. Specific audio data can be selected and a property applied to that audio data. The property is indicated in the audio bar 40 by changing the color of the audio bar 40 to a color associated with that property. For example, the “to do” property may be indicated to the user in blue ink for the corresponding note marks and as a blue band in the audio bar 40 for the corresponding audio data.
  • Properties can be assigned to the audio data portions either before or after the audio data is recorded. If the user knows beforehand that the audio data about to be recorded has a certain property, the user can select that property from a menu, assign it to the audio data and begin recording. Properties may be assigned after recording by selecting the audio and assigning a property to the selected audio. For example, toggling one of the [0037] property buttons 50, a property corresponding to the property button is applied to the current turn in the conversation, indicating that this audio data portion has the given property. The property buttons 50 shown in the user interface 42 of FIG. 3 include a property button 52 which does not assign a property, a topic property button 54, a name property button 56, a quote property button 58, a phone property button 60, a to do property button 62, a Uniform Resource Locator (URL) property button 64, an organization property button 66 and a user defined property button. Properties can be assigned to audio data and note marks that are currently being received or which have been selected by toggling the property button 50 which corresponds to the desired property. The audio data system 10 may also indicate an audio data property, even though there are no associated note marks with that property.
  • While the extent of an ink property is defined by the note marks, the extent of an audio property is not automatically defined. However, since there is a natural segmentation of the audio according to the turn-taking in the conversation, turns can be used to define the extent of the audio property. Audio properties replace the notion of the audio highlighting used in previous systems, since a property such as “important” can be used instead. Further, the user does not have to specify the extent of the property. [0038]
  • In one embodiment of the [0039] audio data system 10 of this invention, the identity of a caller is used to index the audio data. This index can be obtained in several ways. The caller can be identified using a database linking a phone number or a caller id to the caller. If the caller is not calling from his own phone, the audio data system 10 of this invention uses a database of audio data files of known callers, and matches the voice from the incoming caller to an audio data file of a known caller. Any conventional speaker identification system may be used. One example of a speaker identification system is disclosed in U.S. Pat. No. 5,655,058, herein incorporated by reference in its entirety. The user can also enter the caller's identity through any other conventional input method. The audio data can also be indexed automatically according to date and time.
  • The user can also use the [0040] audio data system 10 of this invention to initiate calls using a database of phone numbers and callers, by specifying a person, either directly, by name, or indirectly, by referring to a note page associated with the person or to a recorded call from the person. FIG. 3 shows a call initiation dialog window 70 on top of user interface 42. A user can access the call initiation dialog window 70 by selecting the call initiation button 72 on the user interface 42. The call initiation dialog window 70 includes a keypad 74 that enables the user to manually enter a telephone number. The call initiation dialog window 70 also includes a phonebook button 76 which provides access to a phonebook listing of telephone numbers (not shown). The call initiation dialog window 70 also has a dropdown text entry box 78 which enables a user to place a cursor in the text entry box 78 and enter text and/or numbers using a separate data input device such as a keyboard. Lastly, the call initiation dialog window 70 has a cancel button 80 which enables a user to cancel a call initiation and also an “ok” button 82 which enables a user to initiate a call once the telephone number information has been entered.
  • This invention also provides a system for call screening and/or routing based on speaker identity. A screen mode (not shown) of the [0041] audio data system 10 can be set so that all calls are automatically answered, all calls are automatically recorded without interaction (as in voice mail), or calls are routed in accordance with the identity of the caller. In one embodiment of the screen mode, the audio data system 10 answers the call and sends a message to the caller to state his name and the purpose of the call. If the caller's identity matches a name on an “accept call” list, the user is notified of the call. Otherwise, the call is recorded as voice mail.
  • In addition to properties, the [0042] audio data system 10 of this invention allows keywords to be associated with audio data. This allows keyword-based retrieval of audio data. In contrast to properties, which describe the type of information contained in the audio data portions having those properties, keywords describe the actual content of the audio data. Because keywords reflect more general information, they are assigned to the entire audio data file and not to particular portions of audio data. Keywords are displayed in the margins of a digital page (not shown) and can be entered in several ways. For example, keywords can be selected from a list of previously entered keywords. Alternately, keywords can be typed in using a pop-up keyboard (not shown). Additionally, keywords can be hand-written in a keyword entry box (not shown) and recognized using a standard handwriting recognition system. In the user-interface shown in FIG. 3, there is a drop down text entry box 124 for entering the keywords. Also, note marks can be selected and sent to a handwriting recognizer, and the resulting text entered as a keyword for the audio data associated with the note marks. If the user does not want to immediately interact with the handwriting recognition system, the user can associate the “keyword” property with digital ink to cause delayed recognition.
  • The audio data can be retrieved using attributes associated with the audio data such as caller, time, and topic, using properties and keywords. The [0043] user interface 42 also has query buttons 112 that enable a user to access a context-based query (not shown) by selecting the context-based query button 114 or access a query window 84 as shown in FIG. 4 using the query button 116. The query window 84 includes property check boxes 86 that enable a user to search based upon a designated property. The query window 84 also has a keyword entry box 88 that permits a user to search for audio data and notemarks having a specified keyword. Retrieval by keyword and name returns a list of audio data files with the identified caller and having the desired keyword. Retrieval by property is more specific, since it provides the audio data portions of calls and digital ink groups having the specified property. This is convenient for locating a particular comment in an audio data file. For example, if a comment is marked with the property “important”, that audio data segment containing the comment can be retrieved using a query with the “important” property designated. Segments of audio data files can also be played to both parties during a conversation. Thus, comments from previous conversations can be shared.
  • Results from a search are shown in FIG. 5 in a [0044] search result window 94. The search result window lists the attributes of the retrieved data files in columns. The columns indicate the creation date 96, the date of last modification 98, the type of property as indicated by icon 100 and the keywords 102. The series of icons describe properties associated with the audio data files. For example, the telephone icon represents the property “phone number” and the check box represents the “action item” property. The final column displays keywords associated with the audio data files.
  • Using the indexing mechanisms provided by the digital properties, time stamping and keywords, the audio data system of this invention provides the user with multiple views of audio data stored in a notebook. The views provided by this invention are analogous to the views in a database system. If a query specifies a particular property, only those audio data portions having that property will be displayed. This view of the audio data is formed by newly generated audio data files, each having an original date and containing only the audio data portions with the specified property. The user can also create a view of a notebook by forming a query to request audio data files having specific keywords. The user can define the keywords in the query by browsing an alphabetized list of keywords that have been attached to pages, or by typing or writing the desired keywords. [0045]
  • As shown in FIG. 4, views are indicated by [0046] tabs 48 at the bottom of the notebook page. A name is associated with each tab 48 for easy reference. The user can return to the previous view by selecting the corresponding tab 48. The user can decide if the view should be static or dynamic. When the view is static, the contents of the view remains the same. When the view is dynamic, new items are added to the view as they are recorded by the notebook, provided that these items satisfy the terms of the query for this view.
  • A user can navigate through a notebook using the [0047] navigation buttons 104 in the user interface window 42. Navigation buttons 106 and 108 allow a user to page forward and backward through pages in a notebook. Navigation Button 110 is a new page button that starts data entry on a fresh notebook page.
  • The [0048] user interface 42 is also provided with mode select buttons 118 which enable a user to select either a page mode using a page mode select button 120 or a table of contents mode using the table of contents mode select button 122. The user interface generally includes the features specified and described in co-pending U.S. patent application Ser. No. 08/821,311.
  • While this invention has been described with the specific embodiments outlined above, many alternatives, modifications and variations are apparent to those skilled in the art. Accordingly, the preferred embodiments described above are illustrative and not limiting. Various changes may be made without departing from the spirit and scope of the invention as defined in the following claims. [0049]

Claims (20)

What is claimed is:
1. A digital audio recording system, comprising:
an audio data input device;
a processor in communication with the audio data input device responsive to the input of audio data to associate at least one property to the audio data; and
a storage device in communication with the processor for storing the audio data.
2. The system of claim 1, further comprising an audio data segmentation system that segments the audio data into at least two audio data segments and wherein the processor associates the at least one property to at least one of the audio segments.
3. The system of claim 2, wherein the audio segmentation system comprises a speaker segmentation algorithm.
4. The system of claim 1, further comprising a pen-based interactive display adapted to receive pen-based input, wherein the processor is responsive to the pen-based input to generate at least one mark comprising digital ink for display on the pen-based interactive display, wherein the digital ink comprises the at least one property and the storage device stores the digital ink.
5. The system of claim 1, wherein the audio data input device comprises a microphone.
6. The system of claim 1, further comprising a call screening device that processes the audio data based upon the property.
7. The system of claim 6, further comprising a database comprising at least one property that identifies at least one caller.
8. The system of claim 1, further comprising a display with an audio data timeline.
9. The system of claim 8, wherein the property is indicated to a user with a color on the display of the audio data timeline.
10. The system of claim 8, wherein the audio data timeline is divided into two audio data timelines each corresponding to a speaker in the audio data.
11. The system of claim 1, wherein the property is one of caller id, speaker id, time of recording, time of last modification, action item, name and keyword.
12. The system of claim 1, wherein the processor is responsive to a user inputted query that identifies a property, searches the storage device and retrieves audio data associated with the identified property.
13. The system of claim 1, wherein the audio data is an audio data file and the property is a keyword.
14. The system of claim 13, wherein the storage device stores a plurality of audio data files each having at least one keyword, wherein the processor is responsive to a user inputted query that identifies a keyword, searches the storage device and retrieves the audio data files having the identified keyword.
15. The system of claim 1, further comprising a user interface for designating the property.
16. The system of claim 15, wherein the user interface is a button that allows the user to selectively apply the property to the audio data.
17. The system of claim 1, wherein the audio data input device is adapted to receive audio data from a telephone system.
18. The system of claim 1, wherein the audio data input device is adapted to receive audio data from a network.
19. The system of claim 18, wherein the audio data input device is adapted to receive audio data from an intranet.
20. The system of claim 18, wherein the audio data input device is adapted to receive audio data from the Internet.
US09/058,283 1998-04-10 1998-04-10 System for recording, annotating and indexing audio data Expired - Lifetime US6404856B1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
US09/058,283 US6404856B1 (en) 1998-04-10 1998-04-10 System for recording, annotating and indexing audio data
JP11089695A JPH11313173A (en) 1998-04-10 1999-03-30 Digital sound recording system
EP99302805A EP0949621A3 (en) 1998-04-10 1999-04-12 System for recording, annotating and indexing audio data

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US09/058,283 US6404856B1 (en) 1998-04-10 1998-04-10 System for recording, annotating and indexing audio data

Publications (2)

Publication Number Publication Date
US20020027976A1 true US20020027976A1 (en) 2002-03-07
US6404856B1 US6404856B1 (en) 2002-06-11

Family

ID=22015844

Family Applications (1)

Application Number Title Priority Date Filing Date
US09/058,283 Expired - Lifetime US6404856B1 (en) 1998-04-10 1998-04-10 System for recording, annotating and indexing audio data

Country Status (3)

Country Link
US (1) US6404856B1 (en)
EP (1) EP0949621A3 (en)
JP (1) JPH11313173A (en)

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6741680B2 (en) * 2000-12-22 2004-05-25 International Business Machines Corporation Telephone message system with flexible presentation capability
US20050005016A1 (en) * 2003-03-13 2005-01-06 Fuji Xerox Co., Ltd. User-data relating apparatus with respect to continuous data
US20050243979A1 (en) * 2004-04-30 2005-11-03 Starbuck Bryan T Integrated messaging user interface with message-based logging
US20070297786A1 (en) * 2006-06-22 2007-12-27 Eli Pozniansky Labeling and Sorting Items of Digital Data by Use of Attached Annotations
US20090111437A1 (en) * 2001-08-17 2009-04-30 Elman Boris S Automated conversation recording device and service
US20090228279A1 (en) * 2008-03-07 2009-09-10 Tandem Readers, Llc Recording of an audio performance of media in segments over a communication network
US20090248645A1 (en) * 2008-03-28 2009-10-01 Brother Kogyo Kabushiki Kaisha Device, method and computer readable medium for management of time-series data
US20090254828A1 (en) * 2004-10-26 2009-10-08 Fuji Xerox Co., Ltd. System and method for acquisition and storage of presentations
US20090285379A1 (en) * 2006-03-24 2009-11-19 Realnetworks Asia Pacific Co., Ltd. Method and system for providing ring back tone played at a point selected by user
US20130311178A1 (en) * 2012-05-21 2013-11-21 Lg Electronics Inc. Method and electronic device for easily searching for voice record
US20140365514A1 (en) * 1999-11-12 2014-12-11 Sony Corporation Apparatus and associated methodology for automatically selecting content using stored criteria
US20180129800A1 (en) * 2016-11-10 2018-05-10 Ricoh Company, Ltd. Information processing apparatus, information processing system, and method of processing information

Families Citing this family (40)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7562392B1 (en) * 1999-05-19 2009-07-14 Digimarc Corporation Methods of interacting with audio and ambient music
GB2359918A (en) 2000-03-01 2001-09-05 Sony Uk Ltd Audio and/or video generation apparatus having a metadata generator
US6697796B2 (en) * 2000-01-13 2004-02-24 Agere Systems Inc. Voice clip search
US7263484B1 (en) 2000-03-04 2007-08-28 Georgia Tech Research Corporation Phonetic searching
US7065182B1 (en) * 2000-08-10 2006-06-20 Glenayre Electronics, Inc. Voice mail message repositioning device
KR100853334B1 (en) * 2000-09-19 2008-08-21 후지쯔 가부시끼가이샤 Store system
US20020118798A1 (en) * 2001-02-27 2002-08-29 Christopher Langhart System and method for recording telephone conversations
US7065187B2 (en) * 2002-01-10 2006-06-20 International Business Machines Corporation System and method for annotating voice messages
US7035807B1 (en) * 2002-02-19 2006-04-25 Brittain John W Sound on sound-annotations
US6853710B2 (en) * 2002-07-17 2005-02-08 Timothy M Harris Telephone call messaging device
US8055503B2 (en) 2002-10-18 2011-11-08 Siemens Enterprise Communications, Inc. Methods and apparatus for audio data analysis and data mining using speech recognition
US20040196306A1 (en) * 2003-04-04 2004-10-07 Microsoft Corporation Method and system for flagging digital ink note content
WO2005020209A2 (en) * 2003-08-22 2005-03-03 Ser Solutions, Ic. System for and method of automated quality monitoring
US8103873B2 (en) * 2003-09-05 2012-01-24 Emc Corporation Method and system for processing auditory communications
US7457396B2 (en) * 2003-09-05 2008-11-25 Emc Corporation Automated call management
US8209185B2 (en) * 2003-09-05 2012-06-26 Emc Corporation Interface for management of auditory communications
US7499531B2 (en) * 2003-09-05 2009-03-03 Emc Corporation Method and system for information lifecycle management
US20050069095A1 (en) * 2003-09-25 2005-03-31 International Business Machines Corporation Search capabilities for voicemail messages
US20060004579A1 (en) * 2004-07-01 2006-01-05 Claudatos Christopher H Flexible video surveillance
US7444287B2 (en) * 2004-07-01 2008-10-28 Emc Corporation Efficient monitoring system and method
US7707037B2 (en) * 2004-07-01 2010-04-27 Emc Corporation Archiving of surveillance data
US8180743B2 (en) 2004-07-01 2012-05-15 Emc Corporation Information management
US8244542B2 (en) * 2004-07-01 2012-08-14 Emc Corporation Video surveillance
US9268780B2 (en) * 2004-07-01 2016-02-23 Emc Corporation Content-driven information lifecycle management
US8180742B2 (en) * 2004-07-01 2012-05-15 Emc Corporation Policy-based information management
US8229904B2 (en) * 2004-07-01 2012-07-24 Emc Corporation Storage pools for information management
US8626514B2 (en) * 2004-08-31 2014-01-07 Emc Corporation Interface for management of multiple auditory communications
US20060075347A1 (en) * 2004-10-05 2006-04-06 Rehm Peter H Computerized notetaking system and method
US8204192B1 (en) * 2004-12-30 2012-06-19 Alcatel Lucent Documented call log
GB2424791A (en) * 2005-03-31 2006-10-04 Digits Ltd Integrated telephone handset, computer caller display, voice call recording and automatic redial system
US7620551B2 (en) * 2006-07-20 2009-11-17 Mspot, Inc. Method and apparatus for providing search capability and targeted advertising for audio, image, and video content over the internet
US8311823B2 (en) * 2006-08-31 2012-11-13 Sony Mobile Communications Ab System and method for searching based on audio search criteria
US8239480B2 (en) * 2006-08-31 2012-08-07 Sony Ericsson Mobile Communications Ab Methods of searching using captured portions of digital audio content and additional information separate therefrom and related systems and computer program products
EP2103155A4 (en) * 2006-12-22 2010-08-11 Motorola Inc Method and device for data capture for push over cellular
WO2008150924A1 (en) * 2007-05-29 2008-12-11 Livescribe, Inc. Animation of audio ink
US9142216B1 (en) * 2012-01-30 2015-09-22 Jan Jannink Systems and methods for organizing and analyzing audio content derived from media files
US10002192B2 (en) * 2009-09-21 2018-06-19 Voicebase, Inc. Systems and methods for organizing and analyzing audio content derived from media files
US20110218994A1 (en) * 2010-03-05 2011-09-08 International Business Machines Corporation Keyword automation of video content
KR101844903B1 (en) * 2011-08-31 2018-04-04 삼성전자 주식회사 Providing Method for Data Complex Recording And Portable Device thereof
JP6128146B2 (en) * 2015-02-24 2017-05-17 カシオ計算機株式会社 Voice search device, voice search method and program

Family Cites Families (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5276731A (en) 1991-04-26 1994-01-04 Rolm Company Method and apparatus for handling incoming telephone calls
AU2868092A (en) * 1991-09-30 1993-05-03 Riverrun Technology Method and apparatus for managing information
US5390236A (en) * 1992-03-31 1995-02-14 Klausner Patent Technologies Telephone answering device linking displayed data with recorded audio message
JPH06178296A (en) * 1992-12-09 1994-06-24 Hitachi Ltd Electronic conference system
JP3138370B2 (en) * 1993-09-09 2001-02-26 株式会社日立製作所 Information processing device
JP3185505B2 (en) * 1993-12-24 2001-07-11 株式会社日立製作所 Meeting record creation support device
US5655058A (en) 1994-04-12 1997-08-05 Xerox Corporation Segmentation of audio data for indexing of conversational speech for real-time or postprocessing applications
GB9408042D0 (en) 1994-04-22 1994-06-15 Hewlett Packard Co Device for managing voice data
US5651055A (en) * 1995-06-29 1997-07-22 Lucent Technologies Inc. Digital secretary
US5559875A (en) * 1995-07-31 1996-09-24 Latitude Communications Method and apparatus for recording and retrieval of audio conferences
US5717879A (en) 1995-11-03 1998-02-10 Xerox Corporation System for the capture and replay of temporal data representing collaborative activities
US5835577A (en) * 1996-04-25 1998-11-10 Copytele, Inc. Multi-functional personal telecommunications apparatus
US6034970A (en) * 1996-05-31 2000-03-07 Adaptive Micro Systems, Inc. Intelligent messaging system and method for providing and updating a message using a communication device, such as a large character display
US5970455A (en) 1997-03-20 1999-10-19 Xerox Corporation System for capturing and retrieving audio data and corresponding hand-written notes

Cited By (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140365514A1 (en) * 1999-11-12 2014-12-11 Sony Corporation Apparatus and associated methodology for automatically selecting content using stored criteria
US6741680B2 (en) * 2000-12-22 2004-05-25 International Business Machines Corporation Telephone message system with flexible presentation capability
US8284907B2 (en) * 2001-08-17 2012-10-09 Verizon Laboratories Inc. Automated conversation recording device and service
US20090111437A1 (en) * 2001-08-17 2009-04-30 Elman Boris S Automated conversation recording device and service
US20050005016A1 (en) * 2003-03-13 2005-01-06 Fuji Xerox Co., Ltd. User-data relating apparatus with respect to continuous data
US7350140B2 (en) * 2003-03-13 2008-03-25 Fuji Xerox Co., Ltd. User-data relating apparatus with respect to continuous data
US7512400B2 (en) * 2004-04-30 2009-03-31 Microsoft Corporation Integrated messaging user interface with message-based logging
AU2004279177B2 (en) * 2004-04-30 2009-11-12 Microsoft Technology Licensing, Llc Integrated messaging user interface with message-based logging
US20050243979A1 (en) * 2004-04-30 2005-11-03 Starbuck Bryan T Integrated messaging user interface with message-based logging
WO2005112409A3 (en) * 2004-04-30 2006-05-18 Microsoft Corp Integrated messaging user interface with message-based logging
US20090254828A1 (en) * 2004-10-26 2009-10-08 Fuji Xerox Co., Ltd. System and method for acquisition and storage of presentations
US9875222B2 (en) * 2004-10-26 2018-01-23 Fuji Xerox Co., Ltd. Capturing and storing elements from a video presentation for later retrieval in response to queries
US9596354B2 (en) * 2006-03-24 2017-03-14 Realnetworks Asia Pacific Co., Ltd. Method and system for providing ring back tone played at a point selected by user
US20090285379A1 (en) * 2006-03-24 2009-11-19 Realnetworks Asia Pacific Co., Ltd. Method and system for providing ring back tone played at a point selected by user
US8718256B2 (en) * 2006-03-24 2014-05-06 Realnetworks Asia Pacific Co., Ltd. Method and system for providing ring back tone played at a point selected by user
US20140321332A1 (en) * 2006-03-24 2014-10-30 Realnetworks Asia Pacific Co., Ltd. Method and system for providing ring back tone played at a point selected by user
US8301995B2 (en) * 2006-06-22 2012-10-30 Csr Technology Inc. Labeling and sorting items of digital data by use of attached annotations
US20070297786A1 (en) * 2006-06-22 2007-12-27 Eli Pozniansky Labeling and Sorting Items of Digital Data by Use of Attached Annotations
US20090228279A1 (en) * 2008-03-07 2009-09-10 Tandem Readers, Llc Recording of an audio performance of media in segments over a communication network
US20090248645A1 (en) * 2008-03-28 2009-10-01 Brother Kogyo Kabushiki Kaisha Device, method and computer readable medium for management of time-series data
US20130311178A1 (en) * 2012-05-21 2013-11-21 Lg Electronics Inc. Method and electronic device for easily searching for voice record
US9224397B2 (en) * 2012-05-21 2015-12-29 Lg Electronics Inc. Method and electronic device for easily searching for voice record
US20180129800A1 (en) * 2016-11-10 2018-05-10 Ricoh Company, Ltd. Information processing apparatus, information processing system, and method of processing information
US10558795B2 (en) * 2016-11-10 2020-02-11 Ricoh Company, Ltd. Information processing apparatus, information processing system, and method of processing information
US11017073B2 (en) 2016-11-10 2021-05-25 Ricoh Company, Ltd. Information processing apparatus, information processing system, and method of processing information

Also Published As

Publication number Publication date
EP0949621A2 (en) 1999-10-13
US6404856B1 (en) 2002-06-11
JPH11313173A (en) 1999-11-09
EP0949621A3 (en) 2001-03-14

Similar Documents

Publication Publication Date Title
US6404856B1 (en) System for recording, annotating and indexing audio data
US5970455A (en) System for capturing and retrieving audio data and corresponding hand-written notes
US5457738A (en) Method and system for searching an on-line directory at a telephone station
JP3175399B2 (en) Card data management device
CN102483917B (en) For the order of display text
Whittaker et al. SCANMail: a voicemail interface that makes speech browsable, readable and searchable
US8781080B2 (en) Systems and methods for presenting audio messages
US5943055A (en) Computer interface method and system
US5742736A (en) Device for managing voice data automatically linking marked message segments to corresponding applications
US6411822B1 (en) Communication terminal
US20030026392A1 (en) Method and apparatus for extracting information from voice messages
US20030128820A1 (en) System and method for gisting, browsing and searching voicemail using automatic speech recognition
CN104679727A (en) Search Based User Interface
US7684552B2 (en) Phone batch calling task management system
JP2009522693A (en) Applying metadata to documents and document objects through the software application user interface
US20090164942A1 (en) User interface and communication terminal
US20020077832A1 (en) Computer based integrated text/graphic document analysis
KR100425831B1 (en) Method of stroing data in a personal information terminal
JP2009522695A (en) How to attach metadata to documents and document objects using the operating system user interface
US8792863B2 (en) Method for collecting and storing annotations associated to a voice audio data
JP4352684B2 (en) Information provision processing system, program
JPH0962740A (en) Consultation system
EP1595208A2 (en) Automatic running and ranking of suitable applications based on given user input
US20070136433A1 (en) Intelligent buffer and related applications
US20070282834A1 (en) Database search method and apparatus utilizing variable number of search boxes

Legal Events

Date Code Title Description
AS Assignment

Owner name: FUJI XEROX CO., LTD., JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:WILCOX, LYNN D.;CHIU, PATRICK R.;GWIZDKA, JACEK;REEL/FRAME:009411/0319;SIGNING DATES FROM 19980626 TO 19980807

Owner name: XEROX CORPORATION, CONNECTICUT

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:WILCOX, LYNN D.;CHIU, PATRICK R.;GWIZDKA, JACEK;REEL/FRAME:009411/0319;SIGNING DATES FROM 19980626 TO 19980807

STCF Information on status: patent grant

Free format text: PATENTED CASE

AS Assignment

Owner name: BANK ONE, NA, AS ADMINISTRATIVE AGENT, ILLINOIS

Free format text: SECURITY AGREEMENT;ASSIGNOR:XEROX CORPORATION;REEL/FRAME:013111/0001

Effective date: 20020621

Owner name: BANK ONE, NA, AS ADMINISTRATIVE AGENT,ILLINOIS

Free format text: SECURITY AGREEMENT;ASSIGNOR:XEROX CORPORATION;REEL/FRAME:013111/0001

Effective date: 20020621

AS Assignment

Owner name: JPMORGAN CHASE BANK, AS COLLATERAL AGENT, TEXAS

Free format text: SECURITY AGREEMENT;ASSIGNOR:XEROX CORPORATION;REEL/FRAME:015134/0476

Effective date: 20030625

Owner name: JPMORGAN CHASE BANK, AS COLLATERAL AGENT,TEXAS

Free format text: SECURITY AGREEMENT;ASSIGNOR:XEROX CORPORATION;REEL/FRAME:015134/0476

Effective date: 20030625

FPAY Fee payment

Year of fee payment: 4

FPAY Fee payment

Year of fee payment: 8

FPAY Fee payment

Year of fee payment: 12

AS Assignment

Owner name: XEROX CORPORATION, NEW YORK

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:BANK ONE, NA;REEL/FRAME:034781/0055

Effective date: 20030625

AS Assignment

Owner name: XEROX CORPORATION, NEW YORK

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:034781/0444

Effective date: 20061204

AS Assignment

Owner name: XEROX CORPORATION, CONNECTICUT

Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE ASSIGNEE'S NAME PREVIOUSLY RECORDED AT REEL: 009411 FRAME: 0319. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT;ASSIGNORS:WILCOX, LYNN D;CHIU, PATRICK R;GWIZDKA, JACEK;SIGNING DATES FROM 19980626 TO 19980807;REEL/FRAME:035226/0537

AS Assignment

Owner name: MAJANDRO LLC, TEXAS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:XEROX CORPORATION;REEL/FRAME:053258/0949

Effective date: 20200413

AS Assignment

Owner name: XEROX CORPORATION, CONNECTICUT

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:JPMORGAN CHASE BANK, N.A. AS SUCCESSOR-IN-INTEREST ADMINISTRATIVE AGENT AND COLLATERAL AGENT TO BANK ONE, N.A.;REEL/FRAME:061388/0388

Effective date: 20220822

Owner name: XEROX CORPORATION, CONNECTICUT

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:JPMORGAN CHASE BANK, N.A. AS SUCCESSOR-IN-INTEREST ADMINISTRATIVE AGENT AND COLLATERAL AGENT TO JPMORGAN CHASE BANK;REEL/FRAME:066728/0193

Effective date: 20220822