CN108027952B - Method and electronic device for providing content - Google Patents

Method and electronic device for providing content Download PDF

Info

Publication number
CN108027952B
CN108027952B CN201680053640.6A CN201680053640A CN108027952B CN 108027952 B CN108027952 B CN 108027952B CN 201680053640 A CN201680053640 A CN 201680053640A CN 108027952 B CN108027952 B CN 108027952B
Authority
CN
China
Prior art keywords
electronic device
user
content
output
processor
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201680053640.6A
Other languages
Chinese (zh)
Other versions
CN108027952A (en
Inventor
辛相旻
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Samsung Electronics Co Ltd
Original Assignee
Samsung Electronics Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Samsung Electronics Co Ltd filed Critical Samsung Electronics Co Ltd
Publication of CN108027952A publication Critical patent/CN108027952A/en
Application granted granted Critical
Publication of CN108027952B publication Critical patent/CN108027952B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/165Management of the audio stream, e.g. setting of volume, audio stream path
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • G10L13/02Methods for producing synthetic speech; Speech synthesisers
    • G10L13/033Voice editing, e.g. manipulating the voice of the synthesiser
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • G10L25/63Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for estimating an emotional state
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/03Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters

Abstract

Electronic devices and methods are provided. The electronic device includes: an audio input module configured to receive a user's voice as a speech input; an audio output module configured to output content corresponding to the voice input; and a processor configured to determine an output scheme of the content based on at least one of a speech rate of the speech, a volume of the speech, and a keyword included in the speech, which are obtained from the analysis of the speech input.

Description

Method and electronic device for providing content
Technical Field
The present disclosure relates generally to voice input for electronic devices, and more particularly, to methods and electronic devices for providing content in response to voice input.
Background
Recently, user input interfaces applied to electronic devices have been able to receive user input based on voice input, in addition to user input based on physical manipulations performed by a user (e.g., input through a physical keypad, keyboard, mouse, or touch pad).
An electronic device implementing a speech input interface receives a user's voice as speech input, converts the speech input into an electrical signal, and provides content to the user based on the converted electrical signal.
Disclosure of Invention
Technical problem
An electronic device that supports a voice input interface can provide, for example, content to a user by outputting sound (e.g., outputting speech). However, for content provided in response to voice input, the user does not share an emotional connection with the electronic device because the electronic device provides the content with a uniform speed, a single tone, and a preset volume, regardless of the user's condition while providing the voice input. In addition, since the electronic device does not consider the nuances according to the form of the user's voice, it is difficult for the electronic device to provide contents suitable for the user's condition.
Technical scheme
An aspect of the present disclosure is to address at least the above problems and/or disadvantages and to provide at least the advantages described below. Accordingly, an aspect of the present disclosure is to provide a content providing method of analyzing a speech input of a user's voice and determining an output scheme of content based on various voice characteristics obtained from the result of the analysis, and an electronic device performing the method.
According to an aspect of the present disclosure, an electronic device is provided. The electronic device includes: an audio input module configured to receive a user's voice as a speech input; an audio output module configured to output content corresponding to the voice input; and a processor configured to determine an output scheme of the content based on at least one of a speech rate (speech rate) of the speech, a volume of the speech, or a keyword included in the speech, which is obtained from the analysis of the speech input.
According to another aspect of the present disclosure, a content providing method of an electronic device is provided. The method includes receiving a voice of a user as a voice input, determining an output scheme of content based on at least one of a voice rate of the voice, a volume of the voice, or a keyword included in the voice, which is obtained from an analysis of the voice input, and outputting the content corresponding to the voice input based on the determined output scheme.
According to another aspect of the present disclosure, instructions recorded on a non-transitory computer-readable recording medium and executed by at least one processor are configured to cause the at least one processor to perform a method. The method includes obtaining a voice of a user as a voice input, determining an output scheme of content based on at least one of a voice rate of the voice, a volume of the voice, or a keyword included in the voice obtained from an analysis of the voice input, and outputting the content corresponding to the voice input based on the determined output scheme.
Drawings
The above and other aspects, features, and advantages of certain embodiments of the present disclosure will become more apparent from the description taken in conjunction with the following drawings, in which:
FIGS. 1a and 1b are diagrams illustrating an environment in which an electronic device operates according to an embodiment of the present disclosure;
FIG. 2 is a diagram illustrating an electronic device according to an embodiment of the present disclosure;
FIG. 3 is a block diagram illustrating an electronic device according to an embodiment of the present disclosure;
fig. 4 is a flowchart illustrating a content providing method according to an embodiment of the present disclosure;
fig. 5a is a flowchart illustrating a distance-based content providing method according to an embodiment of the present disclosure;
fig. 5b is a flowchart illustrating a content providing method based on a rate of movement of a user according to an embodiment of the present disclosure;
fig. 6 is a flowchart illustrating a content providing method based on a voice rate of a user's voice according to an embodiment of the present disclosure;
fig. 7 is a flowchart illustrating a keyword-based content providing method according to an embodiment of the present disclosure;
fig. 8 is a flowchart illustrating a content providing method based on an emotional condition of a user according to an embodiment of the present disclosure;
FIG. 9 is a block diagram illustrating an electronic device according to an embodiment of the present disclosure; and
FIG. 10 is a block diagram illustrating program modules according to an embodiment of the disclosure.
Detailed Description
Embodiments of the present disclosure may be described with reference to the accompanying drawings. Thus, those of ordinary skill in the art will recognize that various modifications, equivalents, and/or substitutions can be made to the embodiments described herein without departing from the scope and spirit of the present disclosure. Similar components may be labeled with similar reference numerals with respect to the description of the figures.
Herein, the expressions "have", "may have", "include", "may include" and "may include" indicate the presence of a corresponding feature (e.g., an element such as a value, a function, an operation, or a component), but do not exclude the presence of additional functions.
As used herein, the expressions "A or B", "at least one of A or/and B", "one or more of A or/and B", and the like, may refer to any and all combinations of one or more of the associated listed terms. For example, the terms "a or B," "at least one of a and B," and "at least one of a or B" may refer to the case where at least one a is included, at least one B is included, or both at least one a and at least one B are included.
As used herein, terms such as "first," "second," and the like, may refer to, but do not limit, various elements of embodiments of the disclosure. For example, "first user equipment" and "second user equipment" indicate different user equipment, regardless of the order or priority of the devices. For example, a first element could be referred to as a second element, and similarly, a second element could be referred to as a first element, without departing from the scope of the disclosure.
When an element (e.g., a first element) is referred to as being "operably or communicatively coupled/coupled" or "connected" to another element (e.g., a second element), the element may be directly coupled or connected to the other element or an intervening element (e.g., a third element) may be present. In contrast, when an element (e.g., a first element) is referred to as being "directly coupled" or "directly connected" to another element (e.g., a second element), it is understood that there are no intervening elements (e.g., third elements).
The expression "configured to," as used herein, may have a definition equivalent to "suitable for," "having … …," designed to, "" adapted to, "" made to, "or" capable, "depending on the situation. The term "configured to" is not limited to being defined as "specifically designed" for hardware. Conversely, the expression "a device configured as … …" may indicate that the device is "capable" of operating with another device or other component. For example, a "processor configured to perform A, B, and C" may refer to a dedicated processor (e.g., an embedded processor) for performing the respective operations, or a general-purpose processor (e.g., a Central Processing Unit (CPU) or an application processor) that may perform the respective operations by executing one or more software programs stored in a memory device.
The terminology used herein is used to describe detailed embodiments of the disclosure and is not intended to limit the scope of the disclosure. Terms in the singular may include the plural unless otherwise specified. Unless otherwise defined herein, all terms used herein, including technical or scientific terms, may have the same definition as commonly understood by one of ordinary skill in the art. It will be further understood that terms, which are defined in dictionaries, and are used generally, should also be defined in accordance with their customary definition in the relevant art and should not be interpreted in an idealized or overly formal sense unless expressly so defined herein for embodiments of the present disclosure. In some cases, terms defined in the specification may not be construed in a manner that excludes embodiments of the present disclosure.
An electronic device according to an embodiment of the present disclosure may include at least one of a smart phone, a tablet Personal Computer (PC), a mobile phone, a video phone, an e-book reader, a desktop PC, a laptop PC, a netbook computer, a workstation, a server, a Personal Digital Assistant (PDA), a Portable Multimedia Player (PMP), a Motion Picture Experts Group (MPEG) (MPEG-1 or MPEG-2) Audio Layer 3(MPEG Audio Layer 3, MP3) player, a mobile medical device, a camera, or a wearable device. The wearable device may include at least one of an accessory type (e.g., a watch, ring, bracelet, foot chain, necklace, glasses, contact lens, or head-mounted-device (HMD)), a fabric or garment integrated type (e.g., electronic clothing), a body-attached type (e.g., a skin pad or tattoo), or an implantable type (e.g., implantable circuitry).
According toIn an embodiment of the present disclosure, the electronic device may be a home appliance. The home appliance may include, for example, a Television (TV), a Digital Versatile Disc (DVD) player, a stereo, a refrigerator, an air conditioner, a cleaner, an oven, a microwave oven, a washing machine, an air purifier, a set-top box, a TV box (e.g., Samsung HomeSync)TM、Apple TVTMOr Google TVTM) Game machine (e.g., Xbox)TMAnd PlayStationTM) At least one of an electronic dictionary, an electronic key, a video camera, an electronic photo frame, and the like.
According to embodiments of the present disclosure, the electronic device may include a medical device (e.g., various portable medical measurement devices (e.g., blood glucose monitoring device, cardiac measurement device, blood pressure measurement device, body temperature measurement device, etc.), a Magnetic Resonance Angiography (MRA) device, a Magnetic Resonance Imaging (MRI) device, a Computed Tomography (CT) device, a scanner, and an ultrasound device), a navigation device, a Global Navigation Satellite System (GNSS) receiver, an Event Data Recorder (EDR), a Flight Data Recorder (FDR), a vehicle-mounted infotainment device, an electronic device for a ship (e.g., navigation system and gyrocompass), an avionic device, a security device, a host for a vehicle, an industrial or home robot, a security device, a host for a vehicle, a medical or home robot, and the like), a Magnetic Resonance Imaging (MRI) device, a GNSS receiver, an EDR (GNSS) receiver, an EDR, a Flight Data Recorder (FDR), a vehicle-mounted infotainment device, and an electronic device for a navigation system for a home, At least one of an Automatic Teller Machine (ATM), a point of sale (POS) device, or an Internet of things (IoT) device (e.g., a light bulb, various sensors, an electric or gas meter, a spray device, a fire alarm, a thermostat, a street light, a toaster, exercise equipment, a hot water tank, a heater, a water heater, etc.).
According to an embodiment of the present disclosure, the electronic device may include at least one of furniture or a part of a building/structure, an electronic board, an electronic signature receiving device, a projector, or various measuring instruments (e.g., a water meter, an electric meter, a gas meter, or a wave meter, etc.). The electronic device may be one of the above devices or a combination thereof. The electronic device may be a flexible electronic device. In addition, the electronic device may not be limited to the above-described electronic devices, and may include other electronic devices and new electronic devices according to development of new technology.
Hereinafter, an electronic apparatus according to an embodiment of the present disclosure will be described with reference to the drawings. The term "user" as used herein may relate to a person using an electronic device, or may relate to a device using an electronic device (e.g., a device implementing artificial intelligence).
Fig. 1a and 1b are diagrams illustrating an environment in which an electronic device operates according to an embodiment of the present disclosure.
Referring to fig. 1a, a user 10, who is spaced a certain distance from the electronic device 100 and is at rest, may speak into the electronic device 100. The voice of the user 10 may be provided to the electronic device 100 as a speech input. For example, the user 10 may say "let i know when now" at a moderate volume and at a moderate voice rate, and voice may be provided to the electronic device 100 as speech input. The electronic device 100 may receive voice input through an embedded audio input module (e.g., a microphone) and may generate content corresponding to a result of analyzing the voice input. For example, electronic device 100 may generate content such as "the current time is nine tenths of a morning" in response to a voice input such as "let me know what time now". The electronic device 100 may provide the generated content as sound through an embedded audio output module (e.g., speaker). In this case, because the user 10 provides speech input at a moderate volume and at a moderate voice rate while at rest, the electronic device 100 may provide the content as sound with a moderate output volume level (volume level) and a moderate output speed. Thus, the user 10 may be provided with content corresponding to the voice input.
Referring to FIG. 1b, a user 10 moving away from the electronic device 100 may speak into the electronic device 100. For example, the user 10 may be busy with a preparation job. Thus, for example, the user 10 may speak a voice with a volume greater than usual and a speech rate faster than usual, such as "what time is now? "of speech. The speech may be provided to the electronic device 100 as a speech input.
The electronic device 100 may receive the voice input through the embedded audio input module and may generate content corresponding to a result of analyzing the voice input. For example, electronic device 100 may respond to a response such as "what time is now? "to generate content such as" nine tenths of a day ". The electronic apparatus 100 may provide the generated content as sound through the embedded audio output module. In this case, because the user 10 provides the speech input at a volume greater than usual and at a speech rate faster than usual while the user 10 is moving away from the electronic device 100, the electronic device 100 may provide the content as sound at a relatively greater output volume level and at a relatively faster output speed. Thus, the user 10 may be provided with content corresponding to the voice input.
Fig. 2 is a diagram illustrating an electronic device according to an embodiment of the present disclosure.
Referring to fig. 2, an electronic device according to an embodiment of the present disclosure may be implemented with a dedicated electronic device 100 operating within a house. The application specific electronic device 100 may include various modules (e.g., elements of fig. 3) for implementing embodiments in accordance with the present disclosure, such as a drive system (e.g., drive motors, various types of articulated joints for robots (e.g., bipedal, quadruped robots), wheels, propellers, etc.) capable of providing mobility to the electronic device 100, a camera capable of recognizing a user, an audio input module capable of receiving voice input, and so forth.
Additionally, electronic devices according to embodiments of the present disclosure may be implemented in a form in which the smartphone 201 and docking station 202 are coupled to each other. For example, the smartphone 201 may provide functionality for implementing embodiments of the present disclosure through various modules (e.g., processor, camera, sensors, etc.) embodied therein. In addition, for example, the docking station 202 may include a charging module (and power supply terminal) capable of providing power to the smartphone 201, a drive system (e.g., drive motors, various types of articulated robotic joints, wheels, propellers, etc.) capable of providing mobility, high power speakers, and the like.
The configuration of an electronic device that can be implemented in various ways as described above will be described below with reference to fig. 3. Elements to be described in fig. 3 may be included, for example, in the electronic device 100 of fig. 2 or in the smartphone 201 and/or docking station 202. The contents providing method of the electronic device 100 will be described with reference to fig. 4 to 8.
Fig. 3 is a block diagram illustrating an electronic device according to an embodiment of the present disclosure.
Referring to fig. 3, an electronic device 101 according to an embodiment of the present disclosure includes a bus 110, a processor 120, a memory 130, an audio module 150, a display 160, a communication interface 170, and a distance detection module 180. The electronic device 101 may not include at least one of the above elements, or may further include other elements. For example, the electronic device 101 may include an input/output interface that provides instructions or data to any other element of the electronic device 101, which are input from a user or another external device.
For example, bus 110 may interconnect the above-described elements 110 through 180, and may include circuitry for communicating communications (e.g., control messages and/or data) among the above-described elements.
The processor 120 may include one or more of a Central Processing Unit (CPU), an Application Processor (AP), or a Communication Processor (CP). For example, the processor 120 may perform arithmetic operations or data processing associated with control and/or communication of at least other elements of the electronic device 101. For example, the processor 120 may run a Voice recognition application (e.g., S-Voice) to perform a content providing method according to an embodiment of the present disclosure.
According to an embodiment of the present disclosure, the processor 120 may analyze a voice input received through the audio input module 151 and may output content corresponding to the voice input through the audio output module 152 in various schemes. For example, content may be provided to a user as sound based on Text To Speech (TTS) technology.
According to an embodiment of the present disclosure, the processor 120 may determine the output scheme of the content based on at least one of a voice rate of the user's voice, a volume of the user's voice, and keywords included in the user's voice, which are obtained from the analysis of the speech input. For example, the output scheme may include an output volume level, an output speed, and an output information amount (output amount of information) of content to be provided as sound.
For example, the output volume level of the content may correspond to the volume level when the content is provided as sound by the audio output module 152. For example, the output speed of the content may correspond to the speed when the content is played back as sound by the audio output module 152. For example, the amount of output information of the content may correspond to the amount of information when the content corresponding to the voice input is provided as sound to the user.
For example, with respect to the amount of output information, the content may be classified into detailed content including rich and expanded information and summary content including only the gist of the response corresponding to the voice input (related examples are described later herein). The detailed content and the digest content may be binary sorted. However, embodiments of the present disclosure may not be limited thereto. For example, the content may be classified into several levels from a format in which the content is most accurately described (a format in which the amount of output information is the largest) to a format in which the amount of output information is the smallest and in which the content is most simply described. For example, the processor 120 may vary or adjust the amount of output information by extracting and reconfiguring a portion of the content that most accurately describes the content. Thus, the processor 120 can adaptively generate contents having various amounts of information.
In addition, according to an embodiment of the present disclosure, the processor 120 may adjust an output speed of the content based on an output information amount of the content. For example, as the amount of output information of the content to be output through the audio output module 152 increases, the output speed of the content may also increase under the control of the processor 120. For example, the output speed of the content may be adjusted depending on the change in the above-described abbreviation level of content (abbrevation level).
According to an embodiment of the present disclosure, the processor 120 may determine an output scheme of the corresponding content based on a distance between the user and the electronic device 101. For example, the processor 120 may determine the distance between the user and the electronic device 101 based on a volume of the user's voice obtained through analysis of the speech input or at least one of the distances calculated, estimated, or measured by the distance monitoring module 180. The processor 120 may adjust at least one of an output volume level of the content, an output speed of the content, or an output information amount of the content based on the determined distance between the user and the electronic device 101.
In addition, according to an embodiment of the present disclosure, the processor 120 may determine an output scheme of the corresponding content based on a rate of movement of the user. For example, the processor 120 may determine the rate of movement of the user based on at least one of the volume of the voice obtained through analysis of the speech input, the frequency shift of the speech input (e.g., where doppler effect is used), or the fluctuation of the distance calculated by the distance monitoring module 180. The method for determining the rate of movement of the user may not be limited to the embodiments of the present disclosure described above, and various speech processing techniques for determining the rate of movement of the user may be used. The processor 120 may adjust at least one of an output volume level of the content, an output speed of the content, or an output information amount of the content based on the determined rate of movement of the user.
In addition, according to an embodiment of the present disclosure, the processor 120 may determine an output scheme of the corresponding content based on a voice rate of the user's voice. For example, the processor 120 may adjust at least one of an output speed of the content or an output information amount of the content based on a speech rate of the user's voice obtained through analysis of the speech input.
In addition, according to an embodiment of the present disclosure, the processor 120 may determine an output scheme of the corresponding content based on keywords included in the user's voice. For example, if the analysis of the speech input indicates that the specified keyword is included in the user's voice, the processor 120 may adjust at least one of an output speed of the content, an output volume level of the content, or an output information amount of the content.
In addition, according to an embodiment of the present disclosure, the processor 120 may determine an output scheme of the corresponding content based on an emotional condition determined based on the voice of the user.
In addition, according to an embodiment of the present disclosure, the processor 120 may adjust an amount of output information of a specific content based on whether a user has an interest in the corresponding content. For example, when outputting content corresponding to a voice input, the processor 120 may determine whether the user has an interest in particular content based on whether additional questions associated with the content (i.e., additional voice inputs) were received after the initial voice input, the word frequency of keywords included in the additional questions, and so on.
For example, when it is determined that the user has an interest in a particular content, the processor 120 may provide more detailed information by increasing the amount of output information for the content in which the user has an interest. Conversely, for example, the processor 120 may reduce the amount of output information for content that is determined to be content that the user has no interest in. That is, the processor 120 of the electronic device 101 may adjust the amount of output information of the content based on the interests of the user, the conditions of the user, the context, and the like.
The memory 130 may include volatile and/or non-volatile memory. For example, the memory 130 may store instructions or data associated with at least one other element of the electronic device 101. Memory 130 may store software and/or programs 140 in accordance with embodiments of the present disclosure. Programs 140 include, for example, a kernel 141, middleware 143, an Application Programming Interface (API) 145, and/or an application program (or application) 147. At least a portion of the kernel 141, the middleware 143, or the API 145 may be referred to as an "Operating System (OS)". According to embodiments of the present disclosure, the applications 147 may include a speech recognition application (e.g., S-Voice) capable of analyzing a speech input and generating and/or providing content corresponding to the speech input.
The audio module 150 includes, for example, an audio input module 151 and an audio output module 152. The audio module 150 may include drivers for driving the audio input module 151 and the audio output module 152 and/or an interface for connecting to an external audio device.
For example, the audio input module 151 may be implemented with a microphone or the like. The audio input module 151 may obtain a user's voice as a speech input. The audio output module 152 may be implemented, for example, with a speaker or the like, and may output content corresponding to a voice input as sound.
The display 160 may include, for example, a Liquid Crystal Display (LCD), a light-emitting diode (LED) display, an Organic LED (OLED) display, a Micro Electro Mechanical Systems (MEMS) display, or an electronic paper display. The display 160 may display, for example, various content (e.g., text, images, videos, icons, symbols, etc.) to a user. The display 160 may include a touch screen and may receive, for example, touch, gesture, proximity, or proximity input using an electronic pen or a portion of the user's body.
The communication interface 170 may establish communication between the electronic device 101 and an external device (e.g., the first external electronic device 102, the second external electronic device 104, or the server 106). For example, the communication interface 170 may connect to the network 162 via wireless communication or wired communication to communicate with an external device (e.g., the second external electronic device 104 or the server 106).
The Wireless communication may include, for example, at least one of long-term evolution (LTE), LTE-Advanced (LTE-a), Code Division Multiple Access (CDMA), Wideband CDMA (Wideband CDMA, WCDMA), Universal Mobile Telecommunications System (UMTS), Wireless Broadband (WiBro), Global System for Mobile Communications (GSM), etc., as a cellular communication protocol. Additionally, the wireless communication may include, for example, a local area network 164. The local area network 164 may include, for example, at least one of a Wi-Fi (Wireless Fidelity) network, a bluetooth network, a Near Field Communication (NFC) network, or a Global Navigation Satellite System (GNSS). The GNSS may include at least one of a Global Positioning System (GPS), a global navigation satellite system (Glonass), a beidou navigation satellite system (beidou), or a european global satellite navigation system (galileo) based on an available area, a bandwidth, and the like. Here, "GPS" and "GNSS" may be used interchangeably. The wired communication may include, for example, at least one of a Universal Serial Bus (USB), a High Definition Multimedia Interface (HDMI), a recommended standard 232 (RS-232), a Plain Old Telephone Service (POTS), and the like. The network 162 may include at least one of a telecommunications network, such as a computer network (e.g., a LAN (local area network) or WAN (wide area network)), the internet, or a telephone network.
The distance detection module 180 may include a depth camera like a time-of-flight (TOF) camera, a stereo camera that calculates depth information using triangulation, a Charge Coupled Device (CCD) camera that calculates distance through image processing, and the like. Additionally or alternatively, the distance detection module 180 may include various sensors, such as infrared sensors, RF (radio frequency) sensors, ultrasonic sensors, and the like. The distance detection module 180 may calculate the distance between the user and the electronic device based on image data obtained from a camera or a sensor by using a computing resource of an embedded controller IC (integrated circuit). Alternatively, the processor 120 may calculate the distance between the user and the electronic device based on the image data obtained from the distance detection module 180 by using its computing resources.
Each of the first and second external electronic devices 102 and 104 may be of the same type as or a different type from the electronic device 101. According to embodiments of the present disclosure, the server 106 may include a group of one or more servers. All or a portion of the operations to be performed by the electronic device 101 may be performed by another electronic device or devices (e.g., the first and second external electronic devices 102 and 104 and the server 106), according to embodiments of the present disclosure. When the electronic device 101 automatically or in response to a request runs any function or service, the electronic device 101 may not execute the function or service internally, but alternatively or additionally, the electronic device 101 may request a function associated with the electronic device 101 at least in part at other devices (e.g., the first external electronic device 102, the second external electronic device 104, or the server 106). Other electronic devices (e.g., the first external electronic device 102, the second external electronic device 104, or the server 106) may execute the requested function or additional functions and may provide the execution result to the electronic device 101. The electronic device 101 may provide the requested function or service using the received results, or may additionally process the received results to provide the requested function or service. For this purpose, for example, cloud computing, distributed computing, or client-server computing may be used.
Fig. 4 is a flowchart illustrating a content providing method according to an embodiment of the present disclosure.
Referring to fig. 4, a content providing method according to an embodiment of the present disclosure may include operations 401 to 405. For example, operations 401 through 405 may be performed by the electronic device 101 of fig. 3.
In operation 401, the electronic device 101 receives a user's voice as a speech input through the audio input module 151.
In operation 403, the electronic device 101 obtains at least one of a speech rate of speech uttered by the user, a volume of the speech, and a keyword included in the speech by analyzing the speech input received in operation 401. The electronic device 101 may determine the output scheme of the content based on at least one of a speech rate of speech uttered by the user, a volume of the speech, or keywords included in the speech. For example, the output scheme of the content may include adjusting an output volume level of the content, an output speed of the content, or an output information amount of the content.
In operation 405, the electronic device 101 outputs content corresponding to the received voice input as sound through the audio output module 152 based on the output scheme determined in operation 403.
Fig. 5a is a flowchart illustrating a content providing method according to an embodiment of the present disclosure.
Referring to fig. 5a, a content providing method according to an embodiment of the present disclosure may include operations 501 to 507. For example, operations 501 through 507 may be performed by electronic device 101 of fig. 3.
In operation 501, the electronic device 101 receives a user's voice as a speech input through the audio input module 151. According to an embodiment of the present disclosure, in operation 501, the electronic device 101 may measure a distance between the user and the electronic device 101 through the distance detection module 180.
In operation 503, the processor 120 of the electronic device 101 obtains a volume of voice uttered by the user by analyzing the speech input received in operation 501, and may determine a distance between the user and the electronic device 101 based on the obtained volume of voice. For example, the processor 120 of the electronic device 101 may determine the distance by using a database that is stored in the memory 130 and in which the volume of the user's voice is related to the distance between the user and the electronic device 101.
According to embodiments of the present disclosure, the processor 120 of the electronic device 101 may determine the distance between the user and the electronic device 101 based on the distance derived from the volume of the voice and the distance calculated by the distance detection module 180. Because the electronic device 101 determines the distance based on both the distance derived from the volume of the voice and the distance calculated by the distance detection module 180, the accuracy of the distance measurement may be improved.
In operation 505, the processor 120 of the electronic device 101 determines an output scheme of the content based on the distance between the user and the electronic device 101 determined in operation 503.
According to an embodiment of the present disclosure, the processor 120 of the electronic device 101 may determine the output scheme of the content such that the output volume level of the respective content increases as the distance between the user and the electronic device 101 increases, and such that the output volume level of the respective content decreases as the distance between the user and the electronic device 101 decreases.
[ Table 1]
Distance (D) between user and electronic device Output volume level of content
D<1m 40dB
1m≤D<2m 45dB
... ...
D≥5m 65dB
For example, as shown in table 1, the electronic device 101 may set the output volume level of the content to 40dB when the distance between the user and the electronic device 101 is less than 1m, and may increase the output volume level by 5dB each time the distance increases by 1 m. For example, the corresponding content may be provided to a user remote from the electronic device 101 as sound at a suitable output volume level.
In addition, table 1 is provided merely as an example, but embodiments of the present disclosure are not limited to the example of table 1. For example, the distance between the user and the electronic device 101 and the output volume level of the content according to the distance may be subdivided, and the difference in the output volume level between sections may be diversified.
According to an embodiment of the present disclosure, the processor 120 of the electronic device 101 may determine the output scheme of the content such that the output speed of the content increases as the distance between the user and the electronic device 101 increases, and such that the output speed of the content decreases as the distance between the user and the electronic device 101 decreases.
[ Table 2]
Distance between user and electronic device (D) Output speed of content
D<1m 150WPM
1m≤D<2m 160WPM
2m≤D<3m 170WPM
... ...
D≥5m 200WPM
For example, as shown in table 2, the electronic apparatus 101 may set an output speed of the content to a standard speed (e.g., 150 Words Per Minute (WPM)) when the distance between the user and the electronic apparatus 101 is less than 1m, and may increase the output speed by 10WPM each time the distance increases by 1 m. However, when the distance between the user and the electronic device 101 is at least equal to 5m, the output speed of the content may not be increased above a certain speed (e.g., 200WPM), even if the distance between the user and the electronic device 101 is further increased.
For example, in most cases, a user requesting content with a voice input to the electronic device 101 may want a quick response while away from the electronic device 101. To this end, the corresponding content may be provided to the user as sound at an appropriate output volume level.
In addition, as in table 1, table 2 is provided as an example. For example, the distance between the user and the electronic device 101 and the output speed of the content according to the distance may be set in more detail.
In addition, according to an embodiment of the present disclosure, the processor 120 of the electronic device 101 may determine the output scheme of the content such that the amount of output information of the content decreases as the distance between the user and the electronic device 101 increases, and such that the amount of output information of the content increases as the distance between the user and the electronic device 101 decreases.
[ Table 3]
Distance between user and electronic device (D) Output information amount of content
D<1m Information amount of original content
1m≤D<2m Contents of the first abbreviation level
2m≤D<3m Contents of the second abbreviation level
3m≤D<4m Contents of the third abbreviation level
D≥4m Fourth abbreviation level content (abstract content)
For example, as shown in table 3, when the distance between the user and the electronic apparatus 101 is less than 1m, the output information amount of the content may be set to be the highest. For example, the content having the largest amount of output information may correspond to the most detailed original content corresponding to the voice input. The electronic apparatus 101 can reduce the amount of output information by making the abbreviation level of the original content higher every time the distance increases by 1 m.
For example, corresponding to the original content of the voice input, such as "let me know the weather of today", may correspond to "the weather of 7, 1/2015 is (r) cloudy and then raining (r). The highest temperature was 28 ℃ and the lowest temperature was 18 ℃ (c), the rainfall was 10mm (r). Prepare your umbrella when you go out (v) ". In this case, the contents of the first abbreviation level may include the information of (r), (c), and (c) of the original contents while excluding (c), and the contents of the second abbreviation level may include the information of (r), (c), and (c) while excluding (c) and (c). In addition, the contents of the third abbreviation level may include only information of (r) and (c), and the contents of the fourth abbreviation level (summary contents) may include only information of (c). Thus, the amount of output information of the content may be varied or adjusted by reconfiguring a portion of the original content that most accurately describes the content.
For example, in most cases, a user requesting content with a voice input to the electronic device 101 may want core content corresponding to the voice input rather than detailed content while away from the electronic device 101. To this end, the appropriate amount of output information corresponding content may be provided to the user based on the distance between the user and the electronic device 101.
Additionally, embodiments of the present disclosure are not limited to the examples of table 1, table 2, and table 3. For example, the distance between the user and the electronic device 101 may be set in more detail, and the abbreviation level for each distance interval may be set in various ways.
In operation 507, the electronic device 101 outputs content corresponding to the received voice input as sound through the audio output module 152 based on the output scheme determined in operation 505.
Fig. 5b is a flowchart illustrating a content providing method according to an embodiment of the present disclosure.
Referring to fig. 5b, a content providing method according to an embodiment of the present disclosure may include operations 502 to 508. For example, operations 502 through 508 may be performed by electronic device 101 of fig. 3.
In operation 502, the electronic device 101 receives a user's voice as a speech input through the audio input module 151. According to an embodiment of the present disclosure, in operation 502, the electronic device 101 measures a distance between the user and the electronic device 101 through the distance detection module 180. For example, the distance between the electronic device 101 and the user may be measured in real time according to a particular time period.
In operation 504, the processor 120 of the electronic device 101 determines a rate of movement (change in distance per unit time, i.e., speed (speed) or velocity (velocity)) of the user by analyzing the voice input received in operation 501. For example, the processor 120 may calculate the rate of movement based on fluctuations in the frequency of the speech input, fluctuations in the volume of the voice, etc., or may determine the rate of movement of the user through various speech processes that are well known.
In accordance with embodiments of the present disclosure, the processor 120 of the electronic device 101 may determine the rate of movement of the user by analyzing fluctuations in the distance measured by the distance detection module 180 (e.g., the time derivative of the measured distance). Additionally, the processor 120 may determine the rate of movement of the user taking into account the rate of movement derived from the analysis of the speech input and the rate of movement based on the measured distance. Because the electronic device 101 considers both the rate of movement derived from the speech input and the rate of movement based on the measured distance, the accuracy of the measurement of the rate of movement may be improved.
In operation 506, the processor 120 determines an output scheme of the corresponding content based on the rate of movement of the user determined in operation 504. For example, the processor 120 may adjust at least one of an output volume level of the content, an output speed of the content, or an output information amount of the content based on the determined rate of movement of the user.
According to embodiments of the present disclosure, the processor 120 of the electronic device 101 may adjust the output volume level of the respective content based on the direction of the rate of movement of the user (e.g., the direction in which the user moves closer to or farther away from the electronic device 101) and/or the magnitude of the rate of movement. For example, the output volume level of the content according to the rate of movement of the user may correspond to table 4.
[ Table 4]
Figure BDA0001597896790000151
Figure BDA0001597896790000161
Referring to table 4, while the user is stationary (i.e., the rate of movement of the user is equal to "0"), the electronic device 101 may set the output volume level of the content to 50 decibels and may decrease the output volume level as the speed of the user approaching the electronic device 101 increases. Additionally, as the speed at which the user moves away from the electronic device 101 increases, the electronic device 101 may increase the output volume level. For example, the corresponding content may be provided to the user as sound at an appropriate output volume level based on the rate of movement.
Table 4 is provided as an example. For example, the rate of movement of the user and the output volume level of the content according to the rate of movement may be set in more detail, and the difference in the output volume level between the intervals of the rate of movement may be set differently.
Additionally, in accordance with embodiments of the present disclosure, the processor 120 of the electronic device 101 may adjust the output volume level of the respective content based on the direction of the rate of movement of the user (e.g., the direction in which the user moves closer to or farther away from the electronic device 101) and/or the magnitude of the rate of movement. For example, the output speed of the content according to the rate of movement of the user may correspond to table 5.
[ Table 5]
Rate of movement of user Output speed of content
Approaching with a speed of 5km/h or higher 135WPM
Approaching with a speed of 2km/h or higher and less than 5km/h 140WPM
Approaching with a speed less than 2km/h 145WPM
0km/h (static state) 150WPM
Moving away with a speed of less than 2km/h 160WPM
Moving away with a speed of 2km/h or more and less than 5km/h 170WPM
With 5km/h or more and less thanMoving away at a speed of 8km/h 180WPM
Moving away with a speed of 8km/h or more 190WPM
For example, as shown in table 5, while the user is in a stationary state (i.e., the rate of movement of the user is equal to "0"), the electronic device 101 may set the output speed of the content to a standard speed (e.g., 150WPM), and may decrease the output speed as the speed at which the user approaches the electronic device 101 becomes faster. In addition, the electronic device 101 may increase the output speed as the speed at which the user moves away from the electronic device 101 becomes faster. However, when the speed at which the user approaches the electronic apparatus 101 is at least equal to 5km/h, or when the speed at which the user moves away from the electronic apparatus 101 is at least equal to 8km/h, the previously decreased/increased output speed may be maintained.
For example, in most cases, a user requesting content with voice input from the electronic device 101 may want a quick response while moving quickly away from the electronic device 101. In this case, the user can be provided with the corresponding content at an output speed suitable for his/her moving situation.
In addition, as in table 4, table 5 is provided as an example. For example, the output speed of the user and the output speed of the content according to the distance may be set in more detail.
Additionally, in accordance with embodiments of the present disclosure, the processor 120 of the electronic device 101 may adjust the output volume level of the respective content based on the direction of the rate of movement of the user (e.g., the direction in which the user moves closer to or farther away from the electronic device 101) and/or the magnitude of the rate of movement. For example, the amount of output information of the content according to the rate of movement of the user may correspond to table 6.
[ Table 6]
Rate of movement of user Output information amount of content
Approaching with a speed of 5km/h or higher Information amount of original content
Approaching with a speed of 2km/h or higher and less than 5km/h Information amount of original content
Approaching with a speed less than 2km/h Information amount of original content
0km/h (static state) Information amount of original content
Moving away with a speed of less than 2km/h Contents of the first abbreviation level
Moving away with a speed of 2km/h or more and less than 5km/h Contents of the second abbreviation level
Moving away with a speed of 5km/h or more and less than 8km/h Contents of the third abbreviation level
Moving away with a speed of 8km/h or more Fourth abbreviation level content (abstract content)
Referring to table 6, as the rate of movement of the user becomes faster in the direction in which the user moves away from the electronic device 101, the abbreviation level of the amount of output information may be increased. That is, if the user moves more rapidly, the user may be provided with content having a higher level of abbreviation. In contrast, when the user is at rest, or when the user moves close to the electronic device 101, the user may be provided with original content (in the most detail) that is not abbreviated at all. Thus, the amount of output information of the content may be varied or adjusted by reconfiguring a portion of the original content that most accurately describes the corresponding content.
For example, in most cases, a user requesting content from the electronic device 101 with a voice input while moving away from the electronic device 101 quickly may want core content rather than detailed content. According to an embodiment of the present disclosure, the electronic device 101 may be able to provide the user with content having an output information amount suitable for the user's moving situation.
In addition, as in tables 4 and 5, table 6 is provided as an example. For example, the rate of movement of the user may be set in more detail, and the abbreviated level for each speed interval may be set in various ways.
In operation 508, the electronic device 101 outputs content corresponding to the received voice input as sound through the audio output module 152 based on the output scheme determined in operation 506.
Fig. 6 is a flowchart illustrating a content providing method according to an embodiment of the present disclosure.
Referring to fig. 6, a content providing method according to an embodiment of the present disclosure may include operations 601 to 607. For example, operations 601-607 may be performed by electronic device 101 of fig. 3.
In operation 601, the electronic device 101 receives a user's voice as a speech input through the audio input module 151.
In operation 603, the processor 120 of the electronic device 101 obtains a speech rate of speech uttered by the user by analyzing the speech input received in operation 601. For example, the speech rate of speech may correspond to a recognition speech rate of a word recognized by applying speech-to-text technology (STT) to a speech input.
In operation 605, the processor 120 of the electronic device 101 determines an output scheme of the content based on the speech rate of the user's speech determined in operation 603.
According to an embodiment of the present disclosure, the processor 120 of the electronic device 101 may determine the output scheme of the content such that the output speed of the respective content increases as the speech rate of the speech uttered by the user increases, and such that the output speed of the respective content decreases as the speech rate of the speech decreases.
[ Table 7]
Figure BDA0001597896790000181
For example, as shown in table 7, when the speech rate of the voice uttered by the user is less than 150WPM, the electronic apparatus 101 may set the output speed of the content to 150 WPM. As the speech rate of the speech uttered by the user increases, the electronic device 101 may increase the output speed of the content by 10WPM for every 10WPM interval. However, if the speech rate of the user's uttered speech is faster than a particular speech rate (e.g., 250WPM), the electronic device 101 may not increase the output speed of the content proportionally, even if the speech rate of the user's uttered speech is further increased. In this way, the corresponding content can be provided as sound at a suitable output speed to the user providing the voice input to the electronic apparatus 101.
Table 7 is provided merely as an example, but embodiments of the present disclosure are not limited to the example of table 7. For example, the output speed of the content may be set to continuously increase in proportion to the speech rate of the user's speech. In addition, the interval of the voice rate of the voice of the user and the output speed corresponding to the content of this interval may be set in more detail, and the output speed of each interval may be set differently.
In addition, according to an embodiment of the present disclosure, the processor 120 of the electronic device 101 may determine the output scheme of the content such that the amount of output information of the content decreases as the speech rate of the speech uttered by the user increases, and such that the amount of output information of the content increases as the speech rate of the speech decreases.
[ Table 8]
Speed of user' S voice (S) Output information amount of content
S<170WPM Information amount of original content
170WPM≤S<190WPM Information content of the first abbreviation level
190WPM≤S<210WPM Information content of the second abbreviation level
210WPM≤S<230WPM Information content of the third abbreviation level
S≥230WPM Information content of the fourth abbreviation level
For example, as shown in table 8, when the speech rate of the user's speech is less than 170WPM, the output information amount of the content may be set to be the highest. For example, the content having the largest amount of output information may correspond to the most detailed original content corresponding to the voice input. The electronic apparatus 101 can reduce the amount of output information by making the abbreviation level of the original content higher each time the speech rate of speech increases by 20 WPM.
A user requesting content for the electronic device 101 with a speech rate of fast speech may correspond to a user who wants to be provided core content quickly. In this case, the electronic apparatus 101 may provide the corresponding content with an appropriate amount of output information in response to the user's intention.
In addition, embodiments of the present disclosure are not limited to the examples of table 8. For example, the distance between the user and the electronic device 101 may be set in more detail, and the abbreviation level for each section may be set in various ways.
In operation 607, the electronic device 101 outputs content corresponding to the received voice input as sound through the audio output module 152 based on the output scheme determined in operation 605.
Fig. 7 is a flowchart illustrating a content providing method according to an embodiment of the present disclosure.
Referring to fig. 7, a content providing method according to an embodiment of the present disclosure may include operations 701 to 707. For example, operations 701 through 707 may be performed by electronic device 101 of fig. 3.
In operation 701, the electronic device 101 receives a user's voice as a speech input through the audio input module 151.
In operation 703, the processor 120 of the electronic device 101 detects a specified keyword from the voice uttered by the user by analyzing the speech input received in operation 701. For example, the speech rate of the speech may correspond to a recognition speech rate of a word recognized by analysis of the speech input.
In operation 705, the processor 120 of the electronic device 101 determines an output scheme of the content based on the keyword detected in operation 703.
For example, if the specified first keyword is included in the user's voice, the processor 120 may increase the output speed of the content. If the second keyword is included, the processor 120 may decrease the output speed of the content. In addition, for example, if the specified third keyword is included in the user's voice, the processor 120 may increase the output volume level of the content. If a fourth key is included, processor 120 may decrease the output volume level of the content. In addition, for example, if the specified fifth keyword is included in the user's voice, the processor 120 may increase the amount of output information of the content. If the specified sixth keyword is included, the processor 120 may reduce the amount of output information of the content.
[ Table 9]
Specified keywords Output scheme for content
The first keyword: fast, agile, immediate, etc Increasing output speed of content
The second keyword: slowly, leisurely, etc Reducing output speed of content
The third key word: loud ground, loud ground and the like Increasing content output volume level
The fourth keyword: "quietly", "gently", etc Reducing content output volume level
The fifth keyword: "in detail"Deeply' etc Increasing content output information volume
The sixth keyword: "simply", etc Reducing content output information volume
Table 9 is an example in which the first to sixth keywords correspond to various content output schemes. With the keywords as described above, the corresponding content can be provided as sound to the user providing voice input for the electronic device 101 with an output scheme that is in line with the intention.
The correspondence between the first to sixth keywords and the content output schemes described in table 9 may be an example, but is not limited to the example. For example, the first through sixth keywords, in addition to the proposed keywords, may include a variety of similar terms, and the output scheme may also be set in various ways. According to an embodiment of the present disclosure, the first to sixth keywords may be designated in advance by a user, or may be automatically set by machine learning using pattern recognition. In addition, the first to sixth keywords may be implemented with phrases or sentences, but are not limited to words.
In operation 707, the electronic apparatus 101 outputs content corresponding to the received voice input as sound through the audio output module 152 based on the output scheme determined in operation 705.
Fig. 8 is a flowchart illustrating a content providing method according to an embodiment of the present disclosure
Referring to fig. 8, a content providing method according to an embodiment of the present disclosure may include operations 801 to 807. For example, operations 801 through 807 may be performed by electronic device 101 of fig. 3.
In operation 801, the electronic device 101 receives a user's voice as a speech input through the audio input module 151.
In operation 803, the processor 120 of the electronic device 101 determines an emotional condition of the user by analyzing the speech input received in operation 801. There may be various methods for determining an emotional condition by analyzing the speech input.
For example, the emotional condition of the user may be determined by applying a Support Vector Machine (SVM) pattern classifier to feature vectors based on features extracted from the speech input signal (e.g., energy, formants, tempo, duration, frequency jitter, amplitude flicker, etc.). Generally, emotional conditions can be classified into six emotions, namely, anger, boredom, happiness, neutrality, sadness, and surprise.
In addition to the emotion described above, the emotional condition of the user may be determined according to at least one of a voice rate of the voice of the user, a volume of the voice, and a keyword included in the voice. For example, if the speech rate of the user's speech becomes greater than a particular speech rate, and if the volume of the speech becomes greater than a particular volume, the processor 120 may determine the emotional condition of the user as an "excited" state. The emotional condition may be automatically set through machine learning using pattern recognition. For example, in determining the state of "active" or "inactive (normal)", a specific voice rate of voice and/or a specific volume of voice may be automatically set by data history accumulated over a period of time.
In operation 805, the processor 120 of the electronic device 101 determines an output scheme of the content based on the emotional condition of the user determined in operation 803. For example, the processor 120 may adjust the output speed, output volume level, or output information amount of the most suitable content in response to the emotional condition of the user determined in operation 803.
In operation 807, the electronic device 101 outputs content corresponding to the received voice input as sound through the audio output module 152 based on the output scheme determined in operation 805.
Even though the respective operations described in fig. 4 to 8 are independently described, some or all of the operations may be combined in various ways and may be performed at the same time or at different times according to an embodiment of the present disclosure. Thus, the electronic apparatus 101 can provide the corresponding content in an output scheme most suitable for the condition of the user who performs the voice input.
According to the embodiments of the present disclosure, it is possible to efficiently provide contents corresponding to a voice input by analyzing the voice input of a voice uttered by a user. Further, various effects directly or indirectly understood through the present disclosure may be provided.
Fig. 9 is a block diagram illustrating an electronic device according to an embodiment of the present disclosure.
Referring to fig. 9, the electronic device 901 may include, for example, all or a part of the electronic device 901 illustrated in fig. 9. All or a portion of the electronic device 901 may be included in the electronic device 100 or the smartphone 201 shown in fig. 1a, 1b, or 2. Also, all or a portion of the electronic device 901 may be included in the electronic device 101.
Electronic device 901 includes one or more processors (e.g., AP)910, a communication module 920, a subscriber identification module 924, memory 930, a sensor module 940, an input device 950, a display module 960, an interface 970, an audio module 980, a camera module 991, a power management module 995, a battery 996, an indicator 997, and a motor 998.
The processor 910 may drive an Operating System (OS) or an application to control a plurality of hardware or software components connected to the processor 910, and may process and calculate various data. For example, the processor 910 may be implemented in a System on Chip (SoC). According to an embodiment of the present disclosure, the processor 910 may further include a Graphic Processing Unit (GPU) and/or an image signal processor. The processor 910 may include at least a portion of the components shown in fig. 9 (e.g., the cellular module 921). The processor 910 may load and process instructions or data received from at least one of the other components (e.g., the non-volatile memory), and may store various data at the non-volatile memory.
The communication module 920 may be configured the same as or similar to the communication interface 970 of fig. 9. The communication module 920 includes a cellular module 921, a Wi-Fi module 923, a Bluetooth (BT) module 925, a GNSS module 927 (e.g., a GPS module, a Glonass module, a beidou module, or a galileo module), a Near Field Communication (NFC) module 928, and a Radio Frequency (RF) module 929.
The cellular module 921 may provide voice communication, video communication, character service, internet service, etc. through a communication network. According to an embodiment of the present disclosure, the cellular module 921 may perform authentication and authorization of the electronic device 901 within a communication network using, for example, a Subscriber Identification Module (SIM) 924 (e.g., a SIM card). Cellular module 921 may perform at least a portion of the functionality provided by processor 910. The cellular module 921 may include a Communication Processor (CP).
For example, each of the Wi-Fi module 923, the BT module 925, the GNSS module 927, and the NFC module 928 may include a processor for processing data exchanged by the respective module. According to an embodiment of the present disclosure, at least a portion (e.g., two or more components) of the cellular module 921, the Wi-Fi module 923, the BT module 925, the GPS module 927, and the NFC module 928 may be included within one Integrated Circuit (IC) or IC package.
The RF module 929 may transmit and receive data, for example, communication signals (e.g., RF signals). The RF module 929 may include a transceiver, a Power Amplifier Module (PAM), a frequency filter, a Low Noise Amplifier (LNA), an antenna, and the like. According to an embodiment of the present disclosure, at least one of the cellular module 921, the Wi-Fi module 923, the BT module 925, the GNSS module 927, or the NFC module 928 may transmit and receive an RF signal through a separate RF module.
The subscriber identification module 924 may include unique identification information (e.g., Integrated Circuit Card Identifier (ICCID)) or subscriber information (e.g., Integrated Mobile Subscriber Identity (IMSI)).
Memory 930 (e.g., memory 130 in fig. 3) includes internal memory 932 and external memory 934. For example, the embedded memory 932 may include at least one of a volatile memory (e.g., Dynamic Random Access Memory (DRAM)), a Static RAM (SRAM), or a Synchronous DRAM (SDRAM)), a non-volatile memory (e.g., one-time programmable read only memory (OTPROM), a programmable ROM (read only memory) (programmable ROM, PROM), an Erasable and Programmable ROM (EPROM), an electrically erasable and programmable ROM (electrically erasable and programmable ROM, EEPROM), a mask ROM, a flash ROM, a NAND (NOR) flash memory, a hard disk drive, or a Solid State Drive (SSD).
The external memory 934 may further include a flash memory drive, such as Compact Flash (CF), Secure Digital (SD), Micro secure digital (Micro-SD), Mini secure digital (Mini-SD), extreme digital (xD), multimedia card (MMC), memory stick, and the like. The external memory 934 may be functionally and/or physically connected to the electronic device 901 via various interfaces.
The sensor module 940 may measure, for example, a physical quantity or may detect an operation state of the electronic device 901. The sensor module 940 may convert the measured or detected information into an electrical signal. For example, the sensor module 940 may include at least one of a gesture sensor 940A, a gyroscope sensor 940B, a pressure sensor 940C, a magnetic sensor 940D, an acceleration sensor 940E, a grip sensor 940F, a proximity sensor 940G, a Red Green Blue (RGB) sensor 940H, a living body sensor 940I, a temperature/humidity sensor 940J, an illuminance sensor 940K, or a UV (ultraviolet) sensor 940M. The sensor module 940 may further include, for example, an electronic nose sensor, an Electromyography (EMG) sensor, an electroencephalogram (EEG) sensor, an Electrocardiogram (ECG) sensor, a photoplethysmography (PPG) sensor, an Infrared (IR) sensor, an iris sensor, a fingerprint sensor, and the like. The sensor module 940 may further include a control circuit for controlling at least one or more sensors included therein. According to an embodiment of the present disclosure, the electronic device 901 may further include a processor that is part of the processor 910 or separate from the processor 910 and configured to control the sensor module 940. The processor may control the sensor module 940 while the processor 910 remains in the sleep state.
The input device 950 includes a touch panel 952, a (digital) pen sensor 954, keys 956, and an ultrasonic input unit 958. The touch panel 952 may use at least one of capacitive, resistive, infrared, and ultrasonic detection methods. Also, the touch panel 952 may further include a control circuit. Touch panel 952 may further include a tactile layer to provide a tactile response to the user.
The (digital) pen sensor 954 may be, for example, part of a touch panel, or may include an additional sheet (sheet) for identification. Keys 956 may include, for example, physical buttons, optical keys, keypads, and the like. The ultrasonic input device 958 may detect (or sense) ultrasonic signals generated from the input device through the microphone 988 and may ensure data corresponding to the detected ultrasonic signals.
Display 960 (e.g., display 160 in FIG. 3) includes a panel 962, a hologram device 964, and a projector 966. The faceplate 962 may be configured the same as or similar to the display 160 shown in FIG. 3. For example, the faceplate 962 may be implemented to be flexible, transparent, or wearable. The panel 962 and the touch panel 952 may be integrated into a single module. The hologram device 964 may display a stereoscopic image in space using an optical interference phenomenon. The projector 966 may project light onto a screen to display an image. The screen may be arranged inside or outside the electronic device 901. The display 960 may further include control circuitry for the control panel 962, the hologram device 964, or the projector 966 according to embodiments of the present disclosure.
The interfaces 970 include, for example, a high-definition multimedia interface (HDMI)972, a Universal Serial Bus (USB)974, an optical interface 976, and a D-subminiature (D-sub) 978. Interface 970 may be included, for example, in communications interface 970 shown in fig. 9. The interface 970 may include, for example, a mobile high definition link (MHL) interface, an SD card/multimedia card (MMC) interface, or an infrared data association (IrDA) standard interface.
The audio module 980 may convert sound and electrical signals in both directions. At least a portion of audio module 980 may be included, for example, in input/output interface 950 shown in fig. 9. The audio module 980 may process, for example, sound information input or output through the speaker 982, the receiver 984, the earphone 986, or the microphone 988.
The camera module 991 for photographing a still image or video may include, for example, at least one image sensor (e.g., a front sensor or a rear sensor), a lens, an Image Signal Processor (ISP), or a flash (e.g., an LED or a xenon lamp).
The power management module 995 can manage, for example, the power of the electronic device 901. According to an embodiment of the present disclosure, a Power Management Integrated Circuit (PMIC), a charger IC, or a battery gauge may be included in the power management module 995. The PMIC may have a wired charging method and/or a wireless charging method. The wireless charging method may include, for example, a magnetic resonance method, a magnetic induction method, or an electromagnetic method, and may further include additional circuits such as a coil loop, a resonance circuit, or a rectifier, etc. The battery gauge may measure, for example, the remaining capacity of the battery 996 and its voltage, current, or temperature while the battery is being charged. The battery 996 may include, for example, a rechargeable battery and/or a solar cell.
The indicator 997 may display a particular state of the electronic device 901 or a portion thereof (e.g., the processor 120 in fig. 3), such as a boot state, a message state, a charging state, and so forth. The motor 998 can convert the electrical signal into a motorMechanical vibrations, and the following effects can be produced: vibration, tactile sensation, and the like. A processing device (e.g., GPU) for supporting mobile TV may be included in electronic device 901. The processing device for supporting mobile TV may be based on Digital Multimedia Broadcasting (DMB), Digital Video Broadcasting (DVB), MediaFloTMEtc. to process the media data.
Each of the above elements of the electronic device according to an embodiment of the present disclosure may be configured with one or more components, and names of the elements may be changed according to the type of the electronic device. An electronic device according to an embodiment of the present disclosure may include at least one of the above elements, and some elements may be omitted or other additional elements may be added. In addition, some of the elements of an electronic device according to embodiments of the present disclosure may be combined with each other to form a single entity that performs the functions of the individual elements in the same manner as before the combination.
FIG. 10 is a block diagram illustrating program modules according to an embodiment of the disclosure.
Referring to fig. 10, program modules 1010 (e.g., programs 140 in fig. 3) may include an Operating System (OS) to control resources associated with electronic device 101 in fig. 3, and/or various applications (e.g., application programs 947) driven on the OS, according to embodiments of the present disclosure. The OS may be, for example, AndroidTM、iOSTM、WindowsTM、SymbianTM、TizenTMOr BadaTM
Program modules 1010 may include a kernel 1020, middleware 1030, an Application Programming Interface (API)1060, and/or applications 1070. At least a portion of the program modules 1010 may be preloaded on the electronic device or downloadable from the first external electronic device 102, the second external electronic device 104, the server 106, etc. in fig. 3.
The kernel 1020 (e.g., kernel 141 in fig. 3) includes, for example, a system resource manager 1021 or a device driver 1023. System resource manager 1021 may perform control, allocation, or retrieval of system resources. According to an embodiment of the present disclosure, system resource manager 1021 may include a process management section, a memory management section, or a file system management section. The device driver 1023 may include, for example, a display driver, a camera driver, a bluetooth driver, a general memory driver, a USB driver, a keypad driver, a Wi-Fi driver, an audio driver, or an inter-process communication (ipc) driver.
The middleware 1030 may provide, for example, a function commonly required by the application 1070, or may provide various functions to the application 1070 through the API 1060 to allow the application 1070 to efficiently use the limited system resources of the electronic device. Middleware 1030 (e.g., middleware 43 in fig. 3) includes, for example, runtime library 1035, application manager 1041, window manager 1042, multimedia manager 1043, resource manager 1044, power manager 1045, database manager 1046, packet manager 1047, connectivity manager 1048, notification manager 1049, location manager 1050, graphics manager 1051, and security manager 1052, according to embodiments of the disclosure.
Runtime libraries 1035 may include, for example, library modules used by a compiler to add new functionality through a programming language while application 1070 is running. The runtime library 1035 may perform input/output management, memory management, or capacity for arithmetic functions.
The application manager 1041 may manage, for example, a lifecycle of at least one of the applications 1070. The window manager 1042 may manage GUI (Graphical User Interface) resources used in the screen. The multimedia manager 1043 may recognize necessary formats for playing various media files and may perform encoding or decoding of the media files by using a codec appropriate to the format. Resource manager 1044 may manage resources, such as at least one application storage space, memory, or source code in applications 1070.
The power manager 1045 may operate with, for example, a basic input/output system (BIOS) to manage a battery or power supply, and may provide power supply information for operation of the electronic device. Database manager 1046 may generate, search, or modify a database used in at least one application in program 1070. The package manager 1047 may install or update an application distributed in the form of a package file.
The connectivity manager 1048 may manage, for example, wireless connections such as Wi-Fi or bluetooth. Notification manager 1049 may display or notify events such as arrival messages, commitments, or proximity notifications in a mode that does not disturb the user. The location manager 1050 may manage location information of the electronic device. The graphic manager 1051 may manage graphic effects provided to a user or manage its associated user interface. The security manager 1052 may provide the necessary general security functions for system security or user authentication. According to an embodiment of the present disclosure, when the electronic device 101 includes telephony functionality, the middleware 1030 may further include a telephony manager for managing voice or video call functionality of the electronic device.
Middleware 1030 may include a middleware module that combines the various functionalities of the above-described components. The middleware 1030 may provide a module dedicated to each OS category to provide differentiated functions. Additionally, middleware 1030 may dynamically remove a portion of pre-existing components, or may add new components thereto.
The API 1060 (e.g., API 145 in fig. 3) may be, for example, a set of programming functions and may be provided with a configuration that is variable depending on the OS. For example, when the OS operating in the device is AndroidTMOr iOSTMIt may be permissible to provide one set of APIs per platform. When the OS operating in the device is TizenTMIt may be permissible to provide two or more API sets per platform.
The application 1070 (e.g., the application 147 in fig. 3) may include, for example, one or more applications capable of providing functions for a home 1071, a dialer 1072, an SMS (Short Messaging Service)/MMS (Multimedia Messaging Service) 1073, an Instant Message (IM) 1074, a browser 1075, a camera 1076, an alarm 1077, a contact 1078, a voice dial 1079, an email 1080, a calendar 1081, a media player 1082, an album 1083, and a clock 1084, or for health care (e.g., measuring the amount of exercise or blood glucose level) or environmental information (e.g., barometric pressure, humidity, or temperature).
According to an embodiment of the present disclosure, the applications 1070 may include an information exchange application to support information exchange between the electronic device 101 and the first external electronic device 102, or the second external electronic device 104. The information exchange application may include, for example, a notification relay application for transmitting specific information to the external electronic device, or a device management application for managing the external electronic device.
For example, the information exchange application may include functionality to send notification information generated from other applications (e.g., applications for SMS/MMS, email, healthcare, or environmental information) to the first external electronic device 102 or the second external electronic device 104. Additionally, the information exchange application may receive, for example, notification information from an external electronic device and provide the notification information to the user.
The device management application may manage (e.g., install, delete, or update), for example, at least one function of the first external electronic device 102 or the second external electronic device 104 (e.g., on/off of the external electronic device itself (or a portion of a component) or adjustment of brightness (or resolution) of a display), the first external electronic device 102 or the second external electronic device 104 communicating with an electronic device, an application running in the external electronic device, or a service (e.g., a call service, a message service, etc.) provided from the external electronic device.
According to embodiments of the present disclosure, the applications 1070 may include applications (e.g., healthcare applications) that are arranged according to properties of the first external electronic device 102 or the second external electronic device 104 (e.g., properties of an ambulatory medical device, which is one type of electronic device). The applications 1070 may include applications received from the server 106, the first external electronic device 102, or the second external electronic device 104. The applications 1070 may include preloaded applications or third party applications downloadable from a server. The component title of the program module 1010 according to an embodiment of the present disclosure may be modifiable depending on the kind of the OS.
According to an embodiment of the disclosure, at least a portion of program module 1010 may be implemented by software, firmware, hardware, or a combination of two or more thereof. At least a portion of the program modules 1010 may be implemented (e.g., run) by the processor 910, for example. At least a portion of the program modules 1010 may include, for example, a module, a program, a routine, a set of instructions, or a process for performing one or more functions.
The term "module," as used herein, may refer to, for example, a unit comprising one or more combinations of hardware, software, and firmware. The term "module" may be used interchangeably with the terms "unit," logic block, "" component, "and" circuit. A module may be the smallest unit of an integrated assembly, or may be a part thereof. A module may be the smallest unit or may be part of a unit for performing one or more functions. The modules may be implemented mechanically or electronically. For example, a module may include at least one of an application-specific integrated IC (ASIC) chip, a Field Programmable Gate Array (FPGA), and a Programmable logic device for performing some known or yet to be developed operation.
At least a portion of an apparatus (e.g., modules or functions thereof) or a method (e.g., operations) according to an embodiment of the present disclosure may be implemented, for example, in the form of program modules by instructions stored in a computer-readable storage medium. The instructions, when executed by one or more processors (e.g., processor 120 in fig. 3), may cause the one or more processors to perform functions corresponding to the instructions. The computer-readable storage medium may be, for example, the memory 130 of fig. 3.
The computer-readable recording medium may include a hard disk, a magnetic medium, a floppy disk, a magnetic medium (e.g., a magnetic tape), an optical medium (e.g., a compact disk read only memory (CD-ROM) and a Digital Versatile Disk (DVD), a magneto-optical medium (e.g., a floppy disk), and a hardware device (e.g., a Read Only Memory (ROM), a Random Access Memory (RAM), or a flash memory).
A module or a program module according to an embodiment of the present disclosure may include at least one of the above elements, or a part of the above elements may be omitted, or additional other elements may be further included. Operations performed by modules, program modules, or other elements according to embodiments of the disclosure may be executed sequentially, in parallel, repeatedly, or in a heuristic approach. Also, some operations or portions of operations may be performed in a different order, omitted, or other operations may be added.
While the disclosure has been shown and described with reference to certain embodiments thereof, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the spirit and scope of the disclosure as defined by the appended claims and their equivalents.

Claims (12)

1. An electronic device, comprising:
a microphone;
a speaker, and
a processor configured to:
a speech input of the user's voice is received via the microphone,
determining a rate of movement of the user based on the volume of the voice,
determining an output scheme of the content based on the moving rate, an
Outputting content corresponding to the voice input via the speaker according to the determined output scheme.
2. The electronic device of claim 1, wherein the processor is further configured to adjust an output volume level of the content based on a rate of movement of the user.
3. The electronic device of claim 1, wherein the processor is further configured to adjust an output speed of the content based on a rate of movement of the user.
4. The electronic device of claim 1, wherein the processor is further configured to adjust an amount of output information of the content based on a rate of movement of the user.
5. The electronic device of claim 1, further comprising:
a distance detection module configured to calculate a distance between the user and the electronic device,
wherein the processor is further configured to determine a rate of movement of the user based on fluctuations in volume of the speech and the calculated distance.
6. The electronic device of claim 1, further comprising:
a distance detection module configured to calculate a distance between the user and the electronic device,
wherein the processor is further configured to determine a distance between the user and the electronic device based on the volume of the voice and the calculated distance.
7. The electronic device of claim 1, wherein the processor is configured to adjust an output speed of the content based on a speech rate of the speech.
8. The electronic device of claim 1, wherein the processor is configured to adjust an amount of output information of the content based on a speech rate of the speech.
9. The electronic device of claim 8, wherein the processor is further configured to adjust the amount of output information by extracting and reconfiguring a portion of the content.
10. The electronic device of claim 1, wherein if a specified keyword is included in the speech, the processor is configured to adjust an output speed of the content, an output volume level of the content, or an output information amount of the content.
11. The electronic device of claim 1, wherein the processor is further configured to further determine an emotional condition of the user through analysis of the speech input, and determine an output scheme for the content based on the emotional condition of the user.
12. A content providing method of an electronic device, the method comprising:
receiving a user's voice as a speech input;
determining a rate of movement of the user based on the volume of the voice;
determining an output scheme of the content based on the moving rate; and outputting content corresponding to the speech input based on the determined output scheme.
CN201680053640.6A 2015-09-18 2016-09-09 Method and electronic device for providing content Active CN108027952B (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
KR10-2015-0132488 2015-09-18
KR1020150132488A KR20170034154A (en) 2015-09-18 2015-09-18 Method and electronic device for providing contents
PCT/KR2016/010196 WO2017048000A1 (en) 2015-09-18 2016-09-09 Method and electronic device for providing content

Publications (2)

Publication Number Publication Date
CN108027952A CN108027952A (en) 2018-05-11
CN108027952B true CN108027952B (en) 2021-11-05

Family

ID=58282766

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201680053640.6A Active CN108027952B (en) 2015-09-18 2016-09-09 Method and electronic device for providing content

Country Status (5)

Country Link
US (1) US10062381B2 (en)
EP (1) EP3335188B1 (en)
KR (1) KR20170034154A (en)
CN (1) CN108027952B (en)
WO (1) WO2017048000A1 (en)

Families Citing this family (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP6807031B2 (en) * 2015-06-10 2021-01-06 ソニー株式会社 Signal processor, signal processing method, and program
WO2017130486A1 (en) * 2016-01-28 2017-08-03 ソニー株式会社 Information processing device, information processing method, and program
WO2017167405A1 (en) * 2016-04-01 2017-10-05 Intel Corporation Control and modification of a communication system
US10276149B1 (en) * 2016-12-21 2019-04-30 Amazon Technologies, Inc. Dynamic text-to-speech output
US20180182375A1 (en) * 2016-12-22 2018-06-28 Essential Products, Inc. Method, system, and apparatus for voice and video digital travel companion
WO2018147687A1 (en) 2017-02-10 2018-08-16 Samsung Electronics Co., Ltd. Method and apparatus for managing voice-based interaction in internet of things network system
WO2018222232A1 (en) * 2017-05-31 2018-12-06 Google Llc Providing hands-free data for interactions
US10614794B2 (en) * 2017-06-15 2020-04-07 Lenovo (Singapore) Pte. Ltd. Adjust output characteristic
JP7000924B2 (en) * 2018-03-06 2022-01-19 株式会社Jvcケンウッド Audio content control device, audio content control method, and audio content control program
US11276396B2 (en) * 2018-05-01 2022-03-15 Dell Products, L.P. Handling responses from voice services
CN109215679A (en) * 2018-08-06 2019-01-15 百度在线网络技术(北京)有限公司 Dialogue method and device based on user emotion
NO20181210A1 (en) * 2018-08-31 2020-03-02 Elliptic Laboratories As Voice assistant
CN109299318A (en) * 2018-11-13 2019-02-01 百度在线网络技术(北京)有限公司 Method, apparatus, storage medium and the terminal device that music is recommended
FR3099844B1 (en) * 2019-08-09 2021-07-16 Do You Dream Up Process for automated processing of an automated conversational device by natural language voice exchange, in particular audio rate adaptation process
KR20210020656A (en) * 2019-08-16 2021-02-24 엘지전자 주식회사 Apparatus for voice recognition using artificial intelligence and apparatus for the same
JP7348027B2 (en) * 2019-10-28 2023-09-20 株式会社日立製作所 Dialogue system, dialogue program, and method of controlling the dialogue system
CN111610947B (en) * 2020-05-09 2022-03-29 东风汽车集团有限公司 Vehicle-mounted end conversation volume automatic regulating system
KR102395164B1 (en) 2020-06-26 2022-05-11 카티어스 주식회사 Method and apparatus for providing speech based conversation service

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102710838A (en) * 2012-04-25 2012-10-03 华为技术有限公司 Volume regulation method and device as well as electronic equipment
CN103327168A (en) * 2012-03-21 2013-09-25 Lg电子株式会社 Mobile terminal and control method thereof
CN103984408A (en) * 2013-02-07 2014-08-13 三星电子株式会社 Mobile terminal supporting a voice talk function, and voice talk method
CN104299622A (en) * 2014-09-23 2015-01-21 深圳市金立通信设备有限公司 Audio processing method
CN104618796A (en) * 2015-02-13 2015-05-13 京东方科技集团股份有限公司 Volume adjusting method and display equipment

Family Cites Families (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030167167A1 (en) * 2002-02-26 2003-09-04 Li Gong Intelligent personal assistants
US7881934B2 (en) * 2003-09-12 2011-02-01 Toyota Infotechnology Center Co., Ltd. Method and system for adjusting the voice prompt of an interactive system based upon the user's state
JP2005128884A (en) * 2003-10-24 2005-05-19 Sony Corp Device and method for editing information content
JP2005202076A (en) 2004-01-14 2005-07-28 Sony Corp Uttering control device and method and robot apparatus
JP3924583B2 (en) * 2004-02-03 2007-06-06 松下電器産業株式会社 User adaptive apparatus and control method therefor
US7412378B2 (en) * 2004-04-01 2008-08-12 International Business Machines Corporation Method and system of dynamically adjusting a speech output rate to match a speech input rate
US20060085183A1 (en) * 2004-10-19 2006-04-20 Yogendra Jain System and method for increasing recognition accuracy and modifying the behavior of a device in response to the detection of different levels of speech
JP2008216735A (en) 2007-03-06 2008-09-18 Fujitsu Ltd Reception robot and method of adapting to conversation for reception robot
KR100855563B1 (en) 2007-03-27 2008-09-02 안희달 Language study system of interactive conversation type with a robot
KR101731346B1 (en) * 2010-11-12 2017-04-28 엘지전자 주식회사 Method for providing display image in multimedia device and thereof
US9431980B2 (en) * 2012-01-30 2016-08-30 Echostar Ukraine Llc Apparatus, systems and methods for adjusting output audio volume based on user location
KR101385316B1 (en) 2012-04-03 2014-04-30 주식회사 로보플래닛 System and method for providing conversation service connected with advertisements and contents using robot
US9047054B1 (en) * 2012-12-20 2015-06-02 Audible, Inc. User location-based management of content presentation
US20140278395A1 (en) * 2013-03-12 2014-09-18 Motorola Mobility Llc Method and Apparatus for Determining a Motion Environment Profile to Adapt Voice Recognition Processing
US9134952B2 (en) * 2013-04-03 2015-09-15 Lg Electronics Inc. Terminal and control method thereof
CN104010147B (en) * 2014-04-29 2017-11-07 京东方科技集团股份有限公司 Automatically adjust the method and audio playing apparatus of audio frequency broadcast system volume
US9424841B2 (en) * 2014-10-09 2016-08-23 Google Inc. Hotword detection on multiple devices
US20160379638A1 (en) * 2015-06-26 2016-12-29 Amazon Technologies, Inc. Input speech quality matching

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103327168A (en) * 2012-03-21 2013-09-25 Lg电子株式会社 Mobile terminal and control method thereof
CN102710838A (en) * 2012-04-25 2012-10-03 华为技术有限公司 Volume regulation method and device as well as electronic equipment
CN102710838B (en) * 2012-04-25 2015-01-21 华为技术有限公司 Volume regulation method and device as well as electronic equipment
CN103984408A (en) * 2013-02-07 2014-08-13 三星电子株式会社 Mobile terminal supporting a voice talk function, and voice talk method
CN104299622A (en) * 2014-09-23 2015-01-21 深圳市金立通信设备有限公司 Audio processing method
CN104618796A (en) * 2015-02-13 2015-05-13 京东方科技集团股份有限公司 Volume adjusting method and display equipment

Also Published As

Publication number Publication date
EP3335188B1 (en) 2019-11-06
CN108027952A (en) 2018-05-11
US10062381B2 (en) 2018-08-28
US20170083281A1 (en) 2017-03-23
KR20170034154A (en) 2017-03-28
EP3335188A4 (en) 2018-10-17
WO2017048000A1 (en) 2017-03-23
EP3335188A1 (en) 2018-06-20

Similar Documents

Publication Publication Date Title
CN108027952B (en) Method and electronic device for providing content
US11636861B2 (en) Electronic device and method of operation thereof
US11450315B2 (en) Electronic apparatus and method for operating same
CN108121490B (en) Electronic device, method and server for processing multi-mode input
CN110199350B (en) Method for sensing end of speech and electronic device implementing the method
US10354643B2 (en) Method for recognizing voice signal and electronic device supporting the same
US10593347B2 (en) Method and device for removing noise using neural network model
US10389873B2 (en) Electronic device for outputting message and method for controlling the same
CN108023934B (en) Electronic device and control method thereof
CN108496220B (en) Electronic equipment and voice recognition method thereof
KR102301880B1 (en) Electronic apparatus and method for spoken dialog thereof
EP3593347B1 (en) Method for operating speech recognition service and electronic device supporting the same
US10811008B2 (en) Electronic apparatus for processing user utterance and server
US10034124B2 (en) Electronic apparatus and method for identifying at least one pairing subject in electronic apparatus
US20140358535A1 (en) Method of executing voice recognition of electronic device and electronic device using the same
US9967744B2 (en) Method for providing personal assistant service and electronic device thereof
US10078441B2 (en) Electronic apparatus and method for controlling display displaying content to which effects is applied
US10573317B2 (en) Speech recognition method and device
KR102653450B1 (en) Method for response to input voice of electronic device and electronic device thereof
US10652680B2 (en) Electronic device and method for controlling input and output by electronic device
US11170764B2 (en) Electronic device for processing user utterance
US20180253202A1 (en) Electronic device and method for controlling application thereof
US20180239754A1 (en) Electronic device and method of providing information thereof
US10299034B2 (en) Electronic device and input/output method thereof
EP3157002A1 (en) Electronic device and method for transforming text to speech utilizing super-clustered common acoustic data set for multi-lingual/speaker

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant