WO2022266209A3 - Conversational and environmental transcriptions - Google Patents

Conversational and environmental transcriptions Download PDF

Info

Publication number
WO2022266209A3
WO2022266209A3 PCT/US2022/033607 US2022033607W WO2022266209A3 WO 2022266209 A3 WO2022266209 A3 WO 2022266209A3 US 2022033607 W US2022033607 W US 2022033607W WO 2022266209 A3 WO2022266209 A3 WO 2022266209A3
Authority
WO
WIPO (PCT)
Prior art keywords
content
input
transcriptions
conversation
conversational
Prior art date
Application number
PCT/US2022/033607
Other languages
French (fr)
Other versions
WO2022266209A2 (en
Inventor
Shiraz Akmal
Aaron Mackay BURNS
Brad Kenneth HERMAN
Original Assignee
Apple Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Apple Inc. filed Critical Apple Inc.
Publication of WO2022266209A2 publication Critical patent/WO2022266209A2/en
Publication of WO2022266209A3 publication Critical patent/WO2022266209A3/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/013Eye tracking input arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/10Office automation; Time management
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/18Speech classification or search using natural language modelling
    • G10L15/1822Parsing for meaning understanding
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/167Audio in a user interface, e.g. using voice commands for navigating, audio feedback

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • Human Computer Interaction (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Business, Economics & Management (AREA)
  • Human Resources & Organizations (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Acoustics & Sound (AREA)
  • Strategic Management (AREA)
  • Artificial Intelligence (AREA)
  • Quality & Reliability (AREA)
  • Operations Research (AREA)
  • Marketing (AREA)
  • Tourism & Hospitality (AREA)
  • General Business, Economics & Management (AREA)
  • Economics (AREA)
  • General Health & Medical Sciences (AREA)
  • Databases & Information Systems (AREA)
  • Telephonic Communication Services (AREA)
  • Machine Translation (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

Systems and processes for transcriptions and transcription assistance are provided. For example, a textual representation of a conversation between a user and at least one conversation participant is obtained. Based on the textual representation, content associated with the conversation is identified, wherein the content includes at least one of a first input from the user and a second input from the at least one conversation participant. In response to a determination that the content is associated with predefined content, a portion of the textual representation is identified based on the content. Based on the identified portion, an output responsive to the at least one of the first input and the second input is provided.
PCT/US2022/033607 2021-06-16 2022-06-15 Conversational and environmental transcriptions WO2022266209A2 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US202163211219P 2021-06-16 2021-06-16
US63/211,219 2021-06-16

Publications (2)

Publication Number Publication Date
WO2022266209A2 WO2022266209A2 (en) 2022-12-22
WO2022266209A3 true WO2022266209A3 (en) 2023-01-19

Family

ID=82404357

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2022/033607 WO2022266209A2 (en) 2021-06-16 2022-06-15 Conversational and environmental transcriptions

Country Status (1)

Country Link
WO (1) WO2022266209A2 (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11868736B1 (en) * 2022-11-09 2024-01-09 Moonbeam, Inc. Approaches to deriving and surfacing insights into conversations in virtual environments and systems for accomplishing the same

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100268534A1 (en) * 2009-04-17 2010-10-21 Microsoft Corporation Transcription, archiving and threading of voice communications
US20180075659A1 (en) * 2016-09-13 2018-03-15 Magic Leap, Inc. Sensory eyewear
US20200207358A1 (en) * 2018-06-26 2020-07-02 Eyesight Mobile Technologies Ltd. Contextual driver monitoring system
US20210117214A1 (en) * 2019-10-18 2021-04-22 Facebook, Inc. Generating Proactive Content for Assistant Systems

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3859005A (en) 1973-08-13 1975-01-07 Albert L Huebner Erosion reduction in wet turbines
US4826405A (en) 1985-10-15 1989-05-02 Aeroquip Corporation Fan blade fabrication system
EP1717677B1 (en) 1998-01-26 2015-06-17 Apple Inc. Method and apparatus for integrating manual input
US7218226B2 (en) 2004-03-01 2007-05-15 Apple Inc. Acceleration-based theft detection system for portable electronic devices
US7688306B2 (en) 2000-10-02 2010-03-30 Apple Inc. Methods and apparatuses for operating a portable device based on an accelerometer
US6677932B1 (en) 2001-01-28 2004-01-13 Finger Works, Inc. System and method for recognizing touch typing under limited tactile feedback conditions
US6570557B1 (en) 2001-02-10 2003-05-27 Finger Works, Inc. Multi-touch system and method for emulating modifier keys via fingertip chords
US7657849B2 (en) 2005-12-23 2010-02-02 Apple Inc. Unlocking a device by performing gestures on an unlock image
US10903964B2 (en) 2017-03-24 2021-01-26 Apple Inc. Techniques to enable physical downlink control channel communications
CN110932673A (en) 2018-09-19 2020-03-27 恩智浦美国有限公司 Chopper-stabilized amplifier containing shunt notch filter

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100268534A1 (en) * 2009-04-17 2010-10-21 Microsoft Corporation Transcription, archiving and threading of voice communications
US20180075659A1 (en) * 2016-09-13 2018-03-15 Magic Leap, Inc. Sensory eyewear
US20200207358A1 (en) * 2018-06-26 2020-07-02 Eyesight Mobile Technologies Ltd. Contextual driver monitoring system
US20210117214A1 (en) * 2019-10-18 2021-04-22 Facebook, Inc. Generating Proactive Content for Assistant Systems

Also Published As

Publication number Publication date
WO2022266209A2 (en) 2022-12-22

Similar Documents

Publication Publication Date Title
US9666209B2 (en) Prevention of unintended distribution of audio information
US8595015B2 (en) Audio communication assessment
WO2008067562A3 (en) Multimodal speech recognition system
KR20180084392A (en) Electronic device and operating method thereof
US9560316B1 (en) Indicating sound quality during a conference
EP3002753A1 (en) Speech enhancement method and apparatus for same
CN106687908A8 (en) For adjusting the gesture rapid mode vocally entered
US9940926B2 (en) Rapid speech recognition adaptation using acoustic input
US20080201142A1 (en) Method and apparatus for automication creation of an interactive log based on real-time content
JPS63301998A (en) Voice recognition responder
KR20090054642A (en) Method for recognizing voice, and apparatus for implementing the same
EP4332959A3 (en) Voice to text conversion based on third-party agent content
EP4235648A3 (en) Language model biasing
WO2022266209A3 (en) Conversational and environmental transcriptions
JP6715943B2 (en) Interactive device, interactive device control method, and control program
WO2003005258A3 (en) Method of providing an account information and method of and device for transcribing of dictations
EP3982358A3 (en) Whisper conversion for private conversations
US20190304442A1 (en) Speech translation device, speech translation method, and recording medium therefor
KR20220130739A (en) speech recognition
JP2012073364A (en) Voice interactive device, method, program
US11138390B2 (en) Concurrent reception of multiple user speech input for translation
JP2005283972A (en) Speech recognition method, and information presentation method and information presentation device using the speech recognition method
JP2006251061A (en) Voice dialog apparatus and voice dialog method
US20210233516A1 (en) Agent system, agent server, and computer readable recording medium
KR102308022B1 (en) Apparatus for recognizing call sign and method for the same

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22738264

Country of ref document: EP

Kind code of ref document: A2

NENP Non-entry into the national phase

Ref country code: DE