WO2021129848A1 - 一种声音播放的控制方法及装置 - Google Patents

一种声音播放的控制方法及装置 Download PDF

Info

Publication number
WO2021129848A1
WO2021129848A1 PCT/CN2020/139782 CN2020139782W WO2021129848A1 WO 2021129848 A1 WO2021129848 A1 WO 2021129848A1 CN 2020139782 W CN2020139782 W CN 2020139782W WO 2021129848 A1 WO2021129848 A1 WO 2021129848A1
Authority
WO
WIPO (PCT)
Prior art keywords
gesture
finger
control instruction
control
sound
Prior art date
Application number
PCT/CN2020/139782
Other languages
English (en)
French (fr)
Inventor
荣志超
罗嘉金
李旭
吕毅博
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Priority to EP20907319.6A priority Critical patent/EP4068052A4/en
Publication of WO2021129848A1 publication Critical patent/WO2021129848A1/zh

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/167Audio in a user interface, e.g. using voice commands for navigating, audio feedback
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/165Management of the audio stream, e.g. setting of volume, audio stream path

Definitions

  • the embodiments of the present application relate to the field of communication technologies, and in particular, to a method and device for controlling sound playback.
  • Sound playback devices used on mobile terminals such as music players or mobile phones generally do not have control functions. To control the sound playback of the mobile terminal, it needs to be done through the mobile terminal. In this case, the user needs to take out the mobile terminal and operate the sound playback on the mobile terminal, which is very inconvenient to use.
  • some sound playback devices can physically wake up the voice assistant and control the sound playback device through voice input.
  • Voice input is not convenient for input in many scenarios, and voice input is not suitable for people with language impairments.
  • some sound playback devices have physical touch control, and the user needs to operate the keys in an invisible situation, which is very easy to mis-operate, and the posture is uncomfortable.
  • the existing equipment is not suitable for them to use.
  • the embodiments of the present application provide a method and device for controlling sound playback, in order to realize a more convenient way of controlling sound playback.
  • a sound playback control method including the following steps: sensing a gesture signal; recognizing a gesture according to the gesture signal; determining the control instruction associated with the gesture, the control instruction Used to indicate sound playback operations.
  • the sound playback control can satisfy more application scenarios. The user does not need to touch the electronic device or the sound playback device or component, and the user does not need voice input. Instead, the user can generate control instructions to control the sound playback through gestures, and the application scenarios are more extensive. For example, for some people with language disabilities, when it is inconvenient to input control instructions by voice, gestures can be used to input control instructions.
  • control instructions through light perception there are many environmental constraints on light perception, and it is easily interfered by ambient light signals. Inputting control instructions through gestures can avoid environmental interference, making it more convenient and accurate.
  • the method does not need to input control instructions through physical touch, and the method of inputting control instructions through gestures can avoid the problem of inconvenient operation due to the small size of the headset, and the operation is more convenient.
  • the gesture signal may include a sound wave signal.
  • the gesture signal sensing may be based on the ultrasonic Doppler effect to sense the gesture signal.
  • recognizing a gesture based on the gesture signal can be achieved by the following steps: converting a sound wave signal into an electrical signal; and recognizing the gesture based on the electrical signal.
  • the method further includes: sending the control instruction to a sound playing device; or executing the sound playing operation indicated by the control instruction.
  • gestures can be recognized in the following ways: feature classification and grouping.
  • feature classification and grouping determine the identity or category of each feature, such as fingers, palms, arms, or other objects. Once classified, multiple features can be divided into corresponding feature groups. The positions in the features in a feature group may be adjacent to each other, or the features in a feature group may come from the same finger.
  • the process of grouping may include filtering, and filtering is used to filter out a part of the features that are not a gesture event event. Filtering can improve the accuracy of gesture recognition. Determine the key parameters of the feature group.
  • the key parameters can include, but are not limited to, the distance between the features, the centroid of the feature, the feature rotation, state transition, the movement rate of the feature or the movement direction of the feature, etc.
  • the key parameter of the feature group can be determined by the recognition algorithm carry out. Through the above process, a unique gesture can be recognized.
  • a control device in a second aspect, includes an induction sensor for sensing a gesture signal; a processor for recognizing a gesture based on the gesture signal; and for determining the gesture associated with the gesture
  • the control instruction is used to instruct the sound playing operation.
  • the sound playback control can satisfy more application scenarios.
  • the user does not need to touch the electronic device or the sound playback device or component, and the user does not need voice input, but can generate control instructions to control the sound playback through gestures.
  • the application scenarios are more extensive. For example, for some people with language disabilities, when it is inconvenient to input control instructions by voice, gestures can be used to input control instructions.
  • control instructions through light perception there are many environmental constraints on light perception, and it is easily interfered by ambient light signals. Inputting control instructions through gestures can avoid environmental interference, making it more convenient and accurate.
  • the method does not need to input control instructions through physical touch, and the method of inputting control instructions through gestures can avoid the problem of inconvenient operation due to the small size of the headset, and the operation is more convenient.
  • control device may further include a memory, which is connected to the processor, or the processor includes the memory.
  • the memory is used to store programs, instructions, or data called by the processor, and when the processor executes the instructions or data stored in the memory, it can implement the first aspect or any of the possible design methods described above.
  • the gesture signal may include a sound wave signal.
  • the inductive sensor may be specifically used to sense gesture signals based on the ultrasonic Doppler effect.
  • the processor when recognizing a gesture based on the gesture signal, is specifically configured to convert a sound wave signal into an electrical signal; and recognize the gesture based on the electrical signal.
  • the processor may also be used to send the control instruction to the sound playing device; or, the processor may also be used to execute the sound playing operation indicated by the control instruction.
  • the processor can send control commands in a wired or wireless manner.
  • control device is a headset, and the headset sends a control instruction to a mobile phone player, and the mobile phone player executes the instruction; or the headset executes the instruction (for example, the sound is increased and decreased).
  • the processor can recognize gestures in the following way: the processor performs feature classification and grouping. In the classification process, determine the identity or category of each feature, such as fingers, palms, arms, or other objects. Once classified, multiple features can be divided into corresponding feature groups. The positions in the features in a feature group may be adjacent to each other, or the features in a feature group may come from the same finger.
  • the process of grouping may include filtering, and filtering is used to filter out a part of the features that are not a gesture event event. Filtering can improve the accuracy of gesture recognition.
  • the processor determines the key parameters of the feature group.
  • the key parameters can include, but are not limited to, the distance between the features, the centroid of the feature, the feature rotation, the state transition, the movement rate of the feature or the movement direction of the feature, etc.
  • the key parameters for determining the feature group can be The recognition algorithm is complete. Through the above process, the processor can recognize a unique gesture.
  • a control device in a third aspect, includes a recognition module for recognizing a gesture based on a gesture signal; a determining module for determining the control instruction associated with the gesture, and the control instruction is used to indicate sound Play operation.
  • the sound playback control can satisfy more application scenarios. The user does not need to touch the electronic device or the sound playback device or component, and the user does not need voice input, but can generate control instructions to control the sound playback through gestures.
  • the application scenarios are more extensive. For example, for some people with language disabilities, when it is inconvenient to input control instructions by voice, gestures can be used to input control instructions.
  • control instructions through light perception there are many environmental constraints on light perception, and it is easily interfered by ambient light signals. Inputting control instructions through gestures can avoid environmental interference, making it more convenient and accurate.
  • the method does not need to input control instructions through physical touch, and the method of inputting control instructions through gestures can avoid the problem of inconvenient operation due to the small size of the headset, and the operation is more convenient.
  • the gesture signal includes a sound wave signal.
  • the recognition module when recognizing a gesture based on the gesture signal, is configured to: convert a sound wave signal into an electrical signal; and recognize the gesture based on the electrical signal.
  • the device further includes a sending module, the sending module is used to send the control instruction to the sound playback device; or, the device further includes an execution module, the execution module is used to execute the control The sound playback operation indicated by the instruction.
  • the sending module can send control commands in a wired or wireless manner.
  • control device is a headset, and the headset sends a control instruction to a mobile phone player, and the mobile phone player executes the instruction; or the headset itself executes the instruction (for example, the sound is increased and decreased).
  • the recognition module can recognize gestures in the following ways: performing feature classification and grouping.
  • the classification process determine the identity or category of each feature, such as fingers, palms, arms, or other objects. Once classified, multiple features can be divided into corresponding feature groups. The positions in the features within a feature group may be adjacent to each other or from the same finger.
  • the process of grouping may include filtering, and filtering is used to filter out a part of the features that are not a gesture event event. Filtering can improve the accuracy of gesture recognition.
  • the recognition module determines the key parameters of the feature group.
  • the key parameters include the distance between the features, the centroid of the feature, the feature rotation, the state transition, the movement rate of the feature or the movement direction of the feature, etc.
  • the key parameters of the feature group can be determined by the recognition algorithm. Through the above process, the recognition module can recognize a unique gesture.
  • the following provides several optional association relationships between gestures and control instructions.
  • the control instruction may correspond to starting playback.
  • the other fingers include one or more of index finger, middle finger, ring finger or little finger.
  • the control instruction when the gesture is to open the fist or the thumb is away from other fingers, the control instruction may correspond to stop playing.
  • the other fingers include one or more of index finger, middle finger, ring finger or little finger.
  • the control instruction may correspond to an increase in volume.
  • the finger is any one or more fingers.
  • the control instruction may be volume reduction .
  • the finger is any one or more fingers.
  • the control instruction may correspond to switching to play the next sound file.
  • the control instruction may correspond to switching to play the previous sound file.
  • the control instruction may correspond to adjusting the volume of the left earphone.
  • the control instruction may correspond to adjusting the volume of the right earphone.
  • a chip system in a fourth aspect, includes a processor, an inductive sensor, and optionally a memory, which is used to implement the first aspect or any one of the possible designs in the first aspect.
  • the chip system can be composed of chips, or it can include chips and other discrete devices.
  • the effect of the chip system can refer to the effects that can be achieved by the methods described in the first aspect and each possible design, and will not be repeated.
  • a computer program product containing instructions, which when running on a computer, causes the computer to execute the method described in the first aspect and any possible design of the first aspect.
  • the effects of the computer program product can refer to the effects that can be achieved by the methods described in the first aspect and each possible design, and will not be repeated here.
  • the embodiments of the present application also provide a computer-readable storage medium, including instructions, which when run on a computer, cause the computer to execute as described in the first aspect or any one of the possible designs in the first aspect.
  • the effects of the computer-readable storage medium may refer to the effects that can be achieved by the methods described in the first aspect and each possible design, and will not be repeated.
  • FIG. 1 is one of the schematic diagrams of the structure of the control device in the embodiment of the application;
  • FIG. 2 is one of the schematic diagrams of the structure of the control device and the sound playback device in an embodiment of the application;
  • FIG. 3a is the second schematic diagram of the structure of the control device and the sound playback device in an embodiment of the application
  • FIG. 3b is the third schematic diagram of the architecture of the control device and the sound playback device in the embodiment of the application;
  • FIG. 4 is a schematic diagram of a method for controlling sound playback by a control device in an embodiment of the application
  • Fig. 5 is the second structural diagram of the control device in the embodiment of the application.
  • FIG. 6 is one of the schematic diagrams of gestures in an embodiment of this application.
  • FIG. 7 is the second schematic diagram of gestures in an embodiment of this application.
  • FIG. 8 is the third schematic diagram of gestures in an embodiment of this application.
  • Fig. 9 is a fourth schematic diagram of gestures in an embodiment of this application.
  • FIG. 10 is the fifth schematic diagram of gestures in an embodiment of this application.
  • FIG. 11 is the sixth schematic diagram of gestures in an embodiment of this application.
  • FIG. 12 is the seventh schematic diagram of gestures in an embodiment of this application.
  • FIG. 13 is the eighth schematic diagram of gestures in an embodiment of this application.
  • FIG. 14 is a schematic flowchart of a method for controlling sound playback in an embodiment of the application.
  • the embodiments of the present application provide a method and device for controlling sound playback, in order to realize a more convenient way of controlling sound playback.
  • the method and the device are based on the same or similar technical conception. Since the method and the device have similar principles for solving the problem, the implementation of the device and the method can be referred to each other, and the repetition will not be repeated.
  • the embodiment of the present application provides a control device, which can execute a sound playback control method.
  • the control device may include one or more modules, and each module is used to execute a corresponding method/operation/step/action.
  • the module can be a hardware circuit, software, or hardware circuit combined with software.
  • the module can be integrated in a processor, or it can exist alone physically, or two or more modules can be integrated in one module.
  • control device 100 may include an inductive sensor 101 and a processor 102.
  • the control device 100 may also include a memory 103.
  • Each module can be used to implement the corresponding steps, as described below.
  • the inductive sensor 101 can be used to sense gesture signals.
  • the inductive sensor 101 may be, for example, an acoustic wave sensor based on the Doppler effect, a WIFI-based sensor, a Bluetooth-based sensor, a millimeter wave-based sensor, or a visible light-based sensor.
  • the inductive sensor 101 can also be replaced with a camera. Obtain gesture signals through the camera.
  • the processor 102 may be used for recognizing a gesture according to the gesture signal sensed by the sensing sensor 101; and determining a control instruction associated with the gesture.
  • Sound playback operations refer to operations related to sound playback performed by the sound playback device.
  • the operation related to sound playback may be to turn on sound playback, turn off sound playback, increase the volume, decrease the volume, play the previous sound file (such as a song file), or play the next sound file, etc.
  • the control device 100 controls the sound playback device according to the recognized control instruction associated with the gesture, and the sound playback device performs a corresponding sound playback operation according to the control instruction.
  • the sound playback device may be a speaker, or a music player, etc.
  • the possible structural relationship between the sound playing device and the control device 100 is as follows.
  • control device 100 may include a sound playing device 104.
  • the sound playing device 104 is connected to the processor 102, or the sound playing device 104, the sensor 101, the processor 102, or the memory 103 are connected through a bus.
  • the control device 100 may be a headset, and may be a wired headset or a wireless headset.
  • the sound playback device 104 is a speaker in a headset.
  • the headset can adjust the playback volume of the speaker according to the control instruction associated with the gesture.
  • control device 100 may also be other types of devices, such as a watch, a sports bracelet, or glasses.
  • control device 100 is coupled with the sound playback device 104.
  • the modules in the control device 100 may be integrated in the sound playback device 104.
  • the processor 102 and the memory 103 are in a sound playing device.
  • a possible application scenario is that the inductive sensor 101 is in a headset, and the sound playback device 104, the processor 102, and the memory 103 are integrated in a mobile phone.
  • the inductive sensor 101 in the headset senses the gesture signal, and transmits the gesture signal to the processor 102 in the mobile phone to recognize the gesture.
  • the processor 102 determines the control instruction associated with the gesture according to the recognized gesture, and controls the sound playback device according to the control instruction 104 performs related sound playback operations.
  • control device 100 can also be connected to a sound playback device 104.
  • the connection method may be a wired or wireless connection, such as a Firewire, a universal serial bus (USB), Bluetooth, or radio frequency (RF) connection.
  • the sound playing device 104 may be a communication device such as a mobile phone, a music player, a personal computer (PC), or a tablet computer.
  • the control device 100 controls the music player or mobile phone to perform sound playback operations according to the recognized control instructions associated with the gestures.
  • the memory 103 is used to store a set of programs.
  • the processor 102 recognizes a gesture according to the gesture signal, it can call the program stored in the memory 103 to recognize the gesture.
  • control instructions can also be used to control components on the electronic device or user interface (UI) components, for example, buttons, scroll wheels, and sliding blocks.
  • UI user interface
  • the sensing sensor 101 can sense the approach, distance, speed, distance, and various gestures of an object.
  • the generation of the gesture signal described in the embodiment of the present application is not limited to various gestures made by a human hand, and may also be a gesture realized by the movement of other objects.
  • the other object is a pen, and gestures can be realized by moving the pen closer or farther away, and gestures can also be realized by turning the pen, such as clockwise rotation and counterclockwise rotation of the pen to realize different gestures.
  • the control device and the control method of sound playback will be described in further detail.
  • the inductive sensor 101 recognizes a gesture signal, and the gesture signal may be a sound wave signal.
  • the inductive sensor 101 may also be referred to as a sound wave transceiver or a sound wave transceiver.
  • the inductive sensor 101 emits sound waves, which may be periodically emitted sound waves.
  • the sensing sensor 101 can recognize the changes brought about by the gesture to the sound wave, such as changes in waveform or amplitude. Based on this change, the gesture signal can be recognized. That is to say, the gesture signal is a kind of sound wave signal, including information such as the waveform or amplitude of the sound wave.
  • the inductive sensor 101 transmits the sound wave signal to the processor 102.
  • the processor 102 recognizes the gesture according to the gesture signal.
  • the processor 102 may first perform at least one of the following processing on the acoustic signal: convert the received acoustic signal into an electrical signal; the electrical signal is an analog signal; analog-to-digital conversion (ADC), that is, an analog signal Convert to digital signal; process the digital signal.
  • ADC analog-to-digital conversion
  • the processor 102 may include a signal processing module and other processing modules.
  • the signal processing module implements the above-mentioned processing of the acoustic signal.
  • the processor 102 is connected to the signal processing module, and the signal processing module connected to the processor 102 receives the sound wave signal of the sound wave sensor, and converts the received sound wave signal into an electric signal;
  • the signal is an analog signal;
  • ADC that is, converts an analog signal into a digital signal; processes the digital signal, and transmits the processed signal to the processor 102, which is not limited in this application.
  • the processor 102 may recognize the gesture according to the processed gesture signal. Gestures refer to various gestures made by the user within the sensing range of the sensing sensor 101. As shown in FIG. 4, when the processor 102 recognizes the gesture according to the gesture signal, it can call a program stored in the memory 103 to recognize the gesture. The processor 102 outputs a control instruction corresponding to the gesture according to the recognized gesture. Optionally, the processor 102 controls different elements of the computer system, such as speakers, according to control instructions corresponding to the recognized gestures. Alternatively, the processor 102 controls the electronic device and related components of the electronic device according to the control instruction corresponding to the recognized gesture. In some cases, an external electronic device or an external handheld communication device may need to be connected through a connecting device.
  • the connection device can be a wired or wireless connection, such as FireWire, USB, Bluetooth, RF.
  • the gesture and the control instruction have an association relationship.
  • the association relationship between multiple gestures and multiple control instructions can be stored in advance. Specifically, this association relationship can be stored in the local storage medium of the control device 100, or of course, can also be stored outside the control device 100. In other remote storage media.
  • one gesture can correspond to one control instruction.
  • multiple gestures correspond to one control instruction, and when the control device senses any one of the multiple gestures, the control instruction can be implemented. It is also possible that one gesture corresponds to multiple control instructions. When the control device senses the gesture, multiple control instructions corresponding to the gesture can be realized. For example, one gesture corresponds to switching to the next sound file and turning up the volume. When the gesture is sensed, it is determined to switch to the next sound file and increase the volume of the next sound file.
  • the memory 103 may be a non-volatile memory, such as a hard disk drive (HDD) or a solid-state drive (SSD), etc., or a volatile memory (volatile memory).
  • a non-volatile memory such as a hard disk drive (HDD) or a solid-state drive (SSD), etc.
  • RAM random-access memory
  • the memory is any other medium that can be used to carry or store desired program codes in the form of instructions or data structures and that can be accessed by a computer, but is not limited to this.
  • the memory in the embodiments of the present application may also be a circuit or any other device capable of realizing a storage function for storing program instructions and/or data.
  • the association relationship between the gesture and the control signaling may be stored in the memory 103.
  • the memory 103 may also store a program for gesture-associated control instructions.
  • the program stored in the memory 103 may be a separate program or a part of the operating system.
  • the processor 102 may be a general-purpose processor, a digital signal processor, an application specific integrated circuit, a field programmable gate array or other programmable logic device, a discrete gate or transistor logic device, or a discrete hardware component.
  • the general-purpose processor may be a microprocessor or any conventional processor or the like.
  • the sound playback device 104 may be a component of a computer system and is located in the same device as the control device 100.
  • the control device 100 is placed in a headset, and the sound playback device 104 is a speaker in the headset.
  • the control device 100 is connected to the sound playback device 104.
  • the control device is placed in a headset.
  • the sound playback device 104 is a related component or device in the tablet computer.
  • the control device in the headset controls the control device in the tablet computer.
  • Related components or equipment perform operations related to control instructions.
  • the sound playback device 104 may provide feedback information to the user in the process of performing the operation indicated by the control instruction.
  • the feedback information may be embodied in at least one aspect of audio, tactile or display.
  • the sound playing device 104 increases the playing volume of the music player on the mobile phone according to the control instruction, and an interface for increasing the volume can be displayed on the screen of the mobile phone.
  • any method may be adopted for the method of recognizing a gesture based on a gesture signal.
  • the embodiment of the present application may implement gesture recognition in the following manner.
  • the user performs a gesture posture, the sensor 101 emits sound waves, and receives a sound wave signal of the gesture.
  • the inductive sensor 101 transmits the sound wave signal to the processor 102, and the processor 102 converts the sound wave signal of the gesture into an electrical signal of the gesture. Convert the analog signal of the gesture into the digital signal of the gesture.
  • the digital signal of the gesture can be displayed through the gesture waveform. The gesture is recognized based on the digital signal of the gesture.
  • the processor 102 converts the gesture signal (which may be a digital signal) into a set of features or a list of features.
  • each feature represents an input, for example, a feature represents a gesture.
  • each feature can correspond to some information, including any one or more of the following: identification (ID), x-axis time, y-axis micro-Doppler frequency, or z-axis amplitude. Use this information to express a feature.
  • ID1, x1, y1, and z1 Use this information to express a feature.
  • the first feature is represented by ID1, x1, y1, and z1
  • the second feature is represented by ID2, x2, y2, and z2.
  • the conversion from a gesture signal (such as a digital signal of a gesture) to a feature can use digital signal processing algorithms.
  • the processor 102 performs feature classification and grouping.
  • the classification process determine the identity or category of each feature, such as fingers, palms, arms, or other objects. Once classified, multiple features can be divided into corresponding feature groups. The positions in the features in a feature group may be adjacent to each other, or the features in a feature group may come from the same finger.
  • the process of grouping may also include filtering, which is used to filter out a part of the features that are not the gesture event event. Filtering can improve the accuracy of gesture recognition.
  • the processor 102 determines the key parameters of the feature group.
  • the key parameters may include the distance between the features, the centroid of the feature, the feature rotation, the state transition, the movement rate of the feature or the movement direction of the feature, etc.
  • the key parameter of the feature group can be determined by the recognition algorithm carry out.
  • the processor 102 can recognize a unique gesture.
  • the processor 102 may recognize the human gesture in the following manner.
  • the time series data is converted into a form (waveform) including a right-handed waveform and a left-handed waveform.
  • Extracting features from the waveform wherein the extraction step includes extracting static features from the waveform including the position of the hand, extracting dynamic features from the waveform, including extracting phase and frequency from the right waveform, and extracting the phase and frequency from the left waveform.
  • the position, phase and frequency of one or more hands in the right-hand waveform and the position, phase and frequency of the hands in the left-hand waveform are synthesized into a rule describing the gesture to recognize the gesture.
  • the above description is based on the realization of the modules of the control device 100 in the form of hardware.
  • the modules of the control device 100 may also be implemented in the form of software.
  • control device 500 when the modules of the control device are implemented in the form of software, the control device 500 includes an identification module 501 and a determination module 502.
  • the recognition module 501 is used for recognizing gestures according to gesture signals
  • the determining module 502 is configured to determine the control instruction associated with the gesture, and the control instruction is used to instruct a sound playing operation.
  • the recognition module 501 when recognizing a gesture based on the gesture signal, is configured to: convert a sound wave signal into an electrical signal; and recognize the gesture based on the electrical signal.
  • the recognition module 501 can recognize gestures in the following manner: performing feature classification and grouping. In the classification process, determine the identity or category of each feature, such as fingers, palms, arms, or other objects. Once classified, multiple features can be divided into corresponding feature groups. The positions in the features within a feature group may be adjacent to each other or from the same finger.
  • the process of grouping may include filtering, and filtering is used to filter out a part of the features that are not a gesture event event. Filtering can improve the accuracy of gesture recognition.
  • the recognition module 501 determines the key parameters of the feature group.
  • the key parameters include the distance between the features, the centroid of the feature, the feature rotation, the state transition, the movement rate of the feature or the movement direction of the feature, etc.
  • the key parameters of the feature group can be determined by the recognition algorithm. Through the above process, the recognition module 501 can recognize a unique gesture.
  • the operations performed by the identification module 501 and the determination module 502 may refer to the corresponding operations performed by the processor 102 described above.
  • the control device 500 also includes a sending module 503, which is used to send a control instruction to the sound playback device; or, the device also includes an execution module 504, which is used to execute the sound indicated by the control instruction. Play operation.
  • the sending module can send control commands in a wired or wireless manner.
  • the architectural relationship between the control device 500 and the sound playback device can also refer to the above-mentioned architectural relationship between the control device 100 and the sound playback device 104.
  • control device 500 may further include a storage module 505, and the function of the storage module 505 can refer to the function of the storage 103, which will not be repeated here.
  • control device may also be implemented in the form of a combination of software and hardware.
  • the hardware and software modules in the processor are combined to perform the above operations performed by the processor.
  • the sound playback control can satisfy more application scenarios.
  • the user does not need to touch the electronic device or the sound playback device or component, and the user does not need voice input, but can generate control instructions to control the sound playback through gestures.
  • the application scenarios are more extensive.
  • gestures can be used to input control instructions.
  • Inputting control instructions through light perception there are many environmental constraints on light perception, and it is easily interfered by ambient light signals.
  • Inputting control instructions through gestures can avoid environmental interference, making it more convenient and accurate.
  • the method of inputting control instructions through gestures can avoid the problem of inconvenient operation due to the small size of the headset, and the operation is more convenient.
  • the processor 102 may determine the control instruction corresponding to the gesture through the above-mentioned processing. It can be understood that the embodiments of the present application are not limited to these gestures and corresponding control instructions.
  • the gestures described below can be used in any combination. Each gesture can be a gesture formed by the left hand or a gesture formed by the right hand. If the control device is a headset, the gesture formed by the left hand can be used to control the left ear headset, and the gesture formed by the right hand can control the right ear headset. Of course, the gesture formed by any hand can also control any headset.
  • the gesture is to make a fist, and the control command is to start playing.
  • a fist is when the fingers are bent toward the palm to form a fist.
  • FIG. 6 from a, b, c, d, to e represent the gesture posture of making a fist.
  • the gesture is to open the fist, and the control instruction is to close or stop playing.
  • e, d, c, b, to a represent the gesture posture of the right hand fist opened.
  • the gesture is to approach the thumb and other fingers, and the control instruction is to start playing.
  • the other fingers include one or more of the index finger, the middle finger, the ring finger or the little finger.
  • from a to b is represented as a gesture of the thumb of the right hand approaching the index finger.
  • the thumb and other fingers can be close to each other by keeping the thumb still and other fingers close to the thumb; or by keeping the other fingers still and the thumb close to the other fingers; or by keeping the thumb and other fingers close to each other.
  • This gesture can also replace the thumb with any other type of finger.
  • the gesture is to move the thumb away from other gestures, and the control command is to close or stop playing.
  • the other fingers include one or more of the index finger, the middle finger, the ring finger or the little finger.
  • from b to a is represented as a gesture of keeping the thumb of the right hand away from the index finger.
  • the distance between the thumb and the other fingers can be that the thumb stays still and the other fingers move away from the thumb; it can also be that the other fingers remain still and the thumb moves away from the other fingers; it can also be that the thumb and other fingers are away from each other.
  • This gesture can also replace the thumb with any other type of finger.
  • the fingers can be close to the pinched state, or close to a certain distance without being pinched. Similarly, the fingers can be moved away from the pinched state, or they can be moved away from a certain distance without being pinched.
  • the gesture described in (3) may also be associated with a control instruction to decrease the volume.
  • the gesture described in (4) may also be associated with a control instruction to increase the volume.
  • Finger pinch corresponds to the mute state.
  • the gesture is the upward movement of the finger, and the control instruction is the increase in volume.
  • the finger can be any one or more fingers. As shown in Figure 8, taking the thumb of the right hand as an example, moving in the upward direction of the arrow indicates a gesture of upward movement of the finger.
  • the gesture is the downward movement of the finger, and the control instruction is the volume decrease.
  • the finger can be any one or more fingers. As shown in Fig. 8, taking the thumb of the right hand as an example, moving in the downward direction of the arrow indicates a gesture of downward movement of the finger.
  • the gesture is the forward movement of the finger, and the control instruction is the increase in volume.
  • the finger can be any one or more fingers. As shown in Figure 9, taking the index finger of the left hand as an example, from a to b represents the gesture of moving the finger forward.
  • the gesture is the backward movement of the finger, and the control instruction is the volume decrease.
  • the finger can be any one or more fingers. As shown in Figure 9, taking the index finger of the left hand as an example, the gesture from b to a represents the backward movement of the finger.
  • the gesture is the movement of the finger to the left, and the control instruction is the increase in volume.
  • the finger can be any one or more fingers. As shown in Figure 10, taking the index finger of the left hand as an example, from a to b represents a gesture of moving the finger to the left.
  • the gesture is the movement of the finger to the right, and the control instruction is the volume decrease.
  • the finger can be any one or more fingers. As shown in Figure 10, taking the index finger of the left hand as an example, from b to a represents a gesture of moving the finger to the right.
  • the gesture is a gesture that simulates rotating the physical knob clockwise, and the control command is to increase the volume. As shown in Figure 11, turn clockwise to achieve this gesture.
  • the gesture is implemented, it is assumed that there is a physical knob, and selecting the physical knob by hand reflects the gesture. Of course, the physical knob does not exist.
  • the gesture is a gesture that simulates rotating the physical knob to turn counterclockwise, and the control command is to decrease the volume. Still as shown in Figure 11, turning counterclockwise realizes the gesture.
  • the gesture is to wave the forearm away from the body, and the control instruction is to switch to play the next sound file.
  • from a to b represents a gesture of waving the forearm away from the body.
  • the gesture is waving the forearm toward the body, and the control instruction is to switch to play the previous sound file.
  • from b to a represents a gesture of waving the forearm away from the body.
  • Gesture one is that the left hand is approaching or away from the left ear, and the control command is the left earphone volume adjustment.
  • the left hand close to the left ear corresponds to an increase in the volume of the left earphone
  • the right hand away from the right ear corresponds to a decrease in the volume of the right earphone.
  • Gesture 2 is to move the right hand closer to or away from the right ear, and the control command is to adjust the volume of the right earphone.
  • Gesture 1 and Gesture 2 can be performed at the same time, and the movement directions of Gesture 1 and Gesture 2 can be the same or different.
  • the left hand is close to the left ear, and the right hand is close to the left ear.
  • the left hand is close to the left ear, and the right hand is away from the left ear.
  • gestures are just some examples, and various control commands may not be limited to the above-mentioned gestures.
  • a gesture with a finger close to the control device the control instruction is volume increase
  • a gesture with a finger away from the control device the control instruction is volume decrease.
  • the aforementioned gesture finger can point to any direction.
  • the finger can point to the control device when the gesture is implemented.
  • gesture 1 is associated with control instruction 1
  • gesture 2 is associated with control instruction 2.
  • the embodiment of the present application may also associate gesture 1 with control instruction 2, gesture 2.
  • Associated control instruction 1 For example, when the finger moves upward, the control command is to increase the volume, and when the finger moves downward, the control command is to decrease the volume.
  • the gesture is swinging the forearm away from the body, and the control instruction is switching to play the previous sound file; the gesture is swinging the forearm toward the body, and the control instruction is switching to play the next sound file.
  • Any two or more gestures described above can be used in combination to form a set of gesture control instructions.
  • control device 100 or the control device 500 described in the foregoing method embodiments of the present application may be implemented by chips or integrated circuits.
  • an embodiment of the present application further provides a chip, including a processor, for supporting the control device 100 or the control device 500 to implement the aforementioned functions.
  • the chip is connected to a memory or the chip includes a memory, and the memory is used to store the necessary program instructions and data of the communication device.
  • the embodiment of the present application provides a computer-readable storage medium that stores a computer program, and the computer program includes instructions for executing a method executed by the control device 100 or the control device 500.
  • the embodiment of the present application provides a computer program product containing instructions, which when running on a computer, causes the computer to execute the method executed by the control device 100 or the control device 500.
  • an embodiment of the present application also provides a sound playback control method, which can be executed by the above-mentioned control device.
  • the specific process of this method is as follows.
  • Sense the gesture signal for example, the gesture signal is sensed by the induction sensor in the control device.
  • the induction sensor in the control device For the specific sensing process, please refer to the detailed description in the above device embodiment, which will not be repeated here.
  • S1402. Recognize the gesture according to the gesture signal.
  • the processor in the control device recognizes the gesture according to the gesture signal.
  • the specific recognition process please refer to the detailed description in the foregoing device embodiment, which will not be repeated here.
  • S1403. Determine a control instruction associated with the recognized gesture, where the control instruction is used to instruct a sound playing operation.
  • the processor in the control device determines the control instruction associated with the recognized gesture.
  • control device may also control the sound playback device according to the recognized control instruction associated with the gesture, for example, it may send the control instruction to the sound playback device.
  • the sound playing device executes a related sound playing operation.
  • the control device is a headset, and the headset sends a control instruction to a mobile phone player, and the mobile phone player executes the control instruction.
  • the control device itself can also perform a corresponding sound playback operation according to the recognized control instruction associated with the gesture.
  • the control device is a headset, and the headset itself executes the control command, for example, the sound of the headset speaker is turned up or down.
  • this application can be provided as methods, systems, or computer program products. Therefore, this application may adopt the form of a complete hardware embodiment, a complete software embodiment, or an embodiment combining software and hardware. Moreover, this application may adopt the form of a computer program product implemented on one or more computer-usable storage media (including but not limited to disk storage, CD-ROM, optical storage, etc.) containing computer-usable program codes.
  • computer-usable storage media including but not limited to disk storage, CD-ROM, optical storage, etc.
  • These computer program instructions can also be stored in a computer-readable memory that can guide a computer or other programmable data processing equipment to work in a specific manner, so that the instructions stored in the computer-readable memory produce an article of manufacture including the instruction device.
  • the device implements the functions specified in one process or multiple processes in the flowchart and/or one block or multiple blocks in the block diagram.
  • These computer program instructions can also be loaded on a computer or other programmable data processing equipment, so that a series of operation steps are executed on the computer or other programmable equipment to produce computer-implemented processing, so as to execute on the computer or other programmable equipment.
  • the instructions provide steps for implementing the functions specified in one process or multiple processes in the flowchart and/or one block or multiple blocks in the block diagram.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • General Health & Medical Sciences (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

一种声音播放的控制方法及装置,以期实现更加方便的控制电子设备的声音播放方式。该控制装置包括:感应传感器,用于感测手势信号;处理器,用于根据所述手势信号,识别手势;以及用于确定与所述手势关联的控制指令,所述控制指令用于指示声音播放操作。进一步的,还可以将所述控制指令发送给声音播放设备;或者,执行所述控制指令所指示的声音播放操作。

Description

一种声音播放的控制方法及装置
相关申请的交叉引用
本申请要求在2019年12月26日提交中国专利局、申请号为201911368976.3、申请名称为“一种声音播放的控制方法及装置”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本申请实施例涉及通信技术领域,尤其涉及一种声音播放的控制方法及装置。
背景技术
用于音乐播放器或者手机等移动终端上的声音播放设备通常不具备控制功能。若控制移动终端的声音播放,需要通过移动终端来完成。这种情况下,用户需要将移动终端取出,在移动终端上操作声音播放,使用很不方便。
现有一种可能的实现方式中,一些声音播放装置可以通过物理唤醒语音助手,通过语音输入控制声音播放设备,语音输入在很多场景下不太方便输入,并且语音输入对于语言障碍的人群并不适用。另一种实现方式中,一些声音播放装置具备物理触摸式控制,用户需要在不可视情况下操作按键,非常容易误操作,而且姿势不舒服。特别的,对于一些残障人士,现有的设备并不适合他们使用。
综上,现有的声音播放的控制方法还需要进一步完善。
发明内容
本申请实施例提供一种声音播放的控制方法及装置,以期实现更加方便的控制声音播放的方式。
第一方面,提供一种声音播放的控制方法,所述方法包括以下步骤:感测手势信号;根据所述手势信号,识别手势;确定与所述手势关联的所述控制指令,所述控制指令用于指示声音播放操作。通过本申请实施例提供的控制方法,使得对声音播放控制可以满足更多的应用场景。用户不需要触摸电子设备或声音播放设备或元件,用户也不需要语音输入,而是通过手势就可以产生控制指令来控制声音播放,应用场景更加宽泛。例如,对于一些语言障碍的人群,在不方便通过语音输入控制指令时,可以借用手势来输入控制指令。对于通过光感来输入控制指令,光感的环境限制因素很多,容易受到环境光信号干扰,而通过手势来输入控制指令,可以避免环境的干扰,更加方便且控制精准。该方法不需要通过物理触摸来输入控制指令,通过手势来输入控制指令的方法可以避免耳机尺寸很小操作不便的问题,操作更加方便。
在一个可能的设计中,所述手势信号可以包括声波信号。所述感测手势信号,可以基于超声波多普勒效应,感测手势信号。
在一个可能的设计中,根据所述手势信号,识别手势,可以通过以下步骤实现:将声波信号转换为电信号;根据所述电信号,识别手势。
在一个可能的设计中,所述方法还包括:将所述控制指令发送给声音播放设备;或者,执行所述控制指令所指示的声音播放操作。
在一个可能的设计中,可以通过以下方式来识别手势:进行特征分类和分组。在分类过程中,确定每个特征的身份或类别,例如手指、手掌、手臂或者其它物体。一旦被分类,多个特征就可以被分到相应的特征组。一个特征组内的特征中的位置可能彼此邻近,或者一个特征组内的特征来自相同的手指。可选的,分组的过程可包括过滤,通过过滤来滤除不是手势姿态事件的一部分特征。过滤可以提高手势识别的精确度。确定特征组的关键参数,关键参数可以但不限于包括特征之间的距离、特征的质心、特征旋转、状态转变、特征的运动速率或特征的运动方向等,确定特征组的关键参数可由识别算法完成。通过以上过程,能够识别出唯一手势。
第二方面,提供一种控制装置,该控制装置包括感应传感器,用于感测手势信号;处理器,用于根据所述手势信号,识别手势;以及用于确定与所述手势关联的所述控制指令,所述控制指令用于指示声音播放操作。通过本申请实施例提供的控制装置,使得对声音播放控制可以满足更多的应用场景。用户不需要触摸电子设备或声音播放设备或元件,用户也不需要语音输入,而是通过手势就可以产生控制指令来控制声音播放。应用场景更加宽泛。例如,对于一些语言障碍的人群,在不方便通过语音输入控制指令时,可以借用手势来输入控制指令。对于通过光感来输入控制指令,光感的环境限制因素很多,容易受到环境光信号干扰,而通过手势来输入控制指令,可以避免环境的干扰,更加方便且控制精准。该方法不需要通过物理触摸来输入控制指令,通过手势来输入控制指令的方法可以避免耳机尺寸很小操作不便的问题,操作更加方便。
在一个可能的设计中,该控制装置还可以包括存储器,存储器与处理器连接,或者处理器包括该存储器。存储器用于存储处理器调用的程序、指令或数据,处理器执行所述存储器中存储的指令或数据时,可以实现上述第一方面或任一种可能的设计描述的方法。
在一个可能的设计中,所述手势信号可以包括声波信号。所述感应传感器具体可以用于基于超声波多普勒效应,感测手势信号。
在一个可能的设计中,在根据所述手势信号,识别手势时,所述处理器具体用于将声波信号转换为电信号;根据所述电信号,识别手势。
在一个可能的设计中,所述处理器还可以用于将所述控制指令发送给声音播放设备;或者,所述处理器还可以用于执行所述控制指令所指示的声音播放操作。处理器可以通过有线或无线的方式发送控制指令。
例如,控制装置为耳机,耳机将控制指令发送给手机播放器,手机播放器执行该指令;或者,耳机来执行该指令(例如声音调大调小)。
在一个可能的设计中,处理器可以通过以下方式来识别手势:处理器进行特征分类和分组。在分类过程中,确定每个特征的身份或类别,例如手指、手掌、手臂或者其它物体。一旦被分类,多个特征就可以被分到相应的特征组。一个特征组内的特征中的位置可能彼此邻近,或者一个特征组内的特征来自相同的手指。可选的,分组的过程可包括过滤,通过过滤来滤除不是手势姿态事件的一部分特征。过滤可以提高手势识别的精确度。处理器确定特征组的关键参数,关键参数可以但不限于包括特征之间的距离、特征的质心、特征旋转、状态转变、特征的运动速率或特征的运动方向等,确定特征组的关键参数可由识别算法完成。通过以上过程,处理器能够识别出唯一手势。
第三方面,提供一种控制装置,该装置包括识别模块,用于根据手势信号,识别手势;确定模块,用于确定与所述手势关联的所述控制指令,所述控制指令用于指示声音播放操作。通过本申请实施例提供的控制装置,使得对声音播放控制可以满足更多的应用场景。用户不需要触摸电子设备或声音播放设备或元件,用户也不需要语音输入,而是通过手势就可以产生控制指令来控制声音播放。应用场景更加宽泛。例如,对于一些语言障碍的人群,在不方便通过语音输入控制指令时,可以借用手势来输入控制指令。对于通过光感来输入控制指令,光感的环境限制因素很多,容易受到环境光信号干扰,而通过手势来输入控制指令,可以避免环境的干扰,更加方便且控制精准。该方法不需要通过物理触摸来输入控制指令,通过手势来输入控制指令的方法可以避免耳机尺寸很小操作不便的问题,操作更加方便。
在一个可能的设计中,所述手势信号包括声波信号。
在一个可能的设计中,在根据所述手势信号,识别手势时,所述识别模块用于:将声波信号转换为电信号;根据所述电信号,识别手势。
在一个可能的设计中,该装置还包括发送模块,所述发送模块用于将所述控制指令发送给声音播放设备;或者,该装置还包括执行模块,所述执行模块用于执行所述控制指令所指示的声音播放操作。发送模块可以通过有线或无线的方式发送控制指令。
例如,控制装置为耳机,耳机将控制指令发送给手机播放器,手机播放器执行该指令;或者,耳机自身来执行该指令(例如声音调大调小)。
在一个可能的设计中,识别模块可以通过以下方式来识别手势:进行特征分类和分组。在分类过程中,确定每个特征的身份或类别,例如手指、手掌、手臂或者其它物体。一旦被分类,多个特征就可以被分到相应的特征组。一个特征组内的特征中的位置可能彼此邻近或者来自相同的手指。可选的,分组的过程可包括过滤,通过过滤来滤除不是手势姿态事件的一部分特征。过滤可以提高手势识别的精确度。识别模块确定特征组的关键参数。关键参数包括特征之间的距离、特征的质心、特征旋转、状态转变、特征的运动速率或特征的运动方向等,确定特征组的关键参数可由识别算法完成。通过以上过程,识别模块能够识别出唯一手势。
结合第一方面提供的方法、第二方面和第三方面提供的装置,下面给出几种可选的手势与控制指令的关联关系。
在一个可能的设计中,所述手势为握拳或大拇指与其他手指靠近时,所述控制指令可以对应为开启播放。其中,所述其他手指包括食指、中指、无名指或小拇指中的一个或多个。
在一个可能的设计中,所述手势为拳头张开或大拇指与其他手指远离时,所述控制指令可以对应为停止播放。其中,所述其他手指包括食指、中指、无名指或小拇指中的一个或多个。
在一个可能的设计中,所述手势为手指向上运动、手指向前运动、手指向左运动或手势模拟旋转物理旋钮顺时针转动中的任意一种时,所述控制指令可以对应为音量增大。其中,所述手指为任意一个或多个手指。
在一个可能的设计中,所述手势为手指向下运动、手指向后运动、手指向右运动或手势模拟旋转物理旋钮逆时针转动中的任意一种时,所述控制指令可以为音量减小。其中,所述手指为任意一个或多个手指。
在一个可能的设计中,所述手势为小臂向远离身体方向挥动时,所述控制指令可以对应为切换播放下一个声音文件。
在一个可能的设计中,所述手势为小臂向靠近身体方向挥动时,所述控制指令可以对应为切换播放上一个声音文件。
在一个可能的设计中,所述手势为左手靠近或远离左耳时,所述控制指令可以对应为左耳机音量调整。
在一个可能的设计中,所述手势为右手靠近或远离右耳时,所述控制指令可以对应为右耳机音量调整。
第四方面,提供一种芯片系统,该芯片系统包括处理器,还可以包括感应传感器,可选的还可以包括存储器,用于实现上述第一方面或第一方面中任一种可能的设计中所述的方法。该芯片系统可以由芯片构成,也可以包含芯片和其他分立器件。该芯片系统的效果可以参照第一方面以及各个可能的设计所述的方法所能达到的效果,不再赘述。
第五方面,提供了一种包含指令的计算机程序产品,当其在计算机上运行时,使得计算机执行上述第一方面和第一方面的任一可能的设计中所述的方法。该计算机程序产品的效果可以参照第一方面以及各个可能的设计所述的方法所能达到的效果,不再赘述。
第六方面,本申请实施例中还提供一种计算机可读存储介质,包括指令,当其在计算机上运行时,使得计算机执行如第一方面或第一方面中任一种可能的设计中所述的方法。该计算机可读存储介质的效果可以参照第一方面以及各个可能的设计所述的方法所能达到的效果,不再赘述。
附图说明
图1为本申请实施例中控制装置结构示意图之一;
图2为本申请实施例中控制装置与声音播放设备的架构示意图之一;
图3a为本申请实施例中控制装置与声音播放设备的架构示意图之二;
图3b为本申请实施例中控制装置与声音播放设备的架构示意图之三;
图4为本申请实施例中控制装置对声音播放的控制方法示意图;
图5为本申请实施例中控制装置结构示意图之二;
图6为本申请实施例中手势示意图之一;
图7为本申请实施例中手势示意图之二;
图8为本申请实施例中手势示意图之三;
图9为本申请实施例中手势示意图之四;
图10为本申请实施例中手势示意图之五;
图11为本申请实施例中手势示意图之六;
图12为本申请实施例中手势示意图之七;
图13为本申请实施例中手势示意图之八;
图14为本申请实施例中声音播放的控制方法流程示意图。
具体实施方式
本申请实施例提供一种声音播放的控制方法及装置,以期实现更加方便的控制声音播 放的方式。其中,方法和装置是基于相同或相似技术构思的,由于方法及装置解决问题的原理相似,因此装置与方法的实施可以相互参见,重复之处不再赘述。
本申请实施例的描述中,“和/或”,描述关联对象的关联关系,表示可以存在三种关系,例如,A和/或B,可以表示:单独存在A,同时存在A和B,单独存在B这三种情况。字符“/”一般表示前后关联对象是一种“或”的关系。本申请中所涉及的至少一个是指一个或多个;多个,是指两个或两个以上。
另外,需要理解的是,在本申请的描述中,“第一”、“第二”、“第三”等词汇,仅用于区分描述的目的,而不能理解为指示或暗示相对重要性,也不能理解为指示或暗示顺序。在本说明书中描述的参考“一个实施例”或“一些实施例”等意味着在本申请的一个或多个实施例中包括结合该实施例描述的特定特征、结构或特点。由此,在本说明书中的不同之处出现的语句“在一个实施例中”、“在一些实施例中”、“在其他一些实施例中”、“在另外一些实施例中”等不是必然都参考相同的实施例,而是意味着“一个或多个但不是所有的实施例”,除非是以其他方式另外特别强调。术语“包括”、“包含”、“具有”及它们的变形都意味着“包括但不限于”,除非是以其他方式另外特别强调。
下面将结合附图,对本申请实施例进行详细描述。
本申请实施例提供一种控制装置,该控制装置可以执行声音播放的控制方法。该控制装置中可以包括一个或多个模块,每一个模块用于执行对应的方法/操作/步骤/动作。该模块可以是硬件电路,也可是软件,也可以是硬件电路结合软件。模块可以集成在一个处理器中,也可以是单独物理存在,也可以两个或两个以上模块集成在一个模块中。
若该控制装置中的模块采用硬件的形式实现,示例性的,如图1所示,控制装置100中可以包括:感应传感器101和处理器102。该控制装置100还可以包括存储器103。各个模块可以用于实现相应的步骤,具体如下所述。
感应传感器101可以用于感测手势信号。感应传感器101例如可以是基于多普勒效应的声波传感器、基于WIFI的传感器、基于蓝牙的传感器、基于毫米波的传感器或者基于可见光的传感器。感应传感器101还可以替换为摄像头。通过摄像头获取手势信号。
处理器102,可以用于根据感应传感器101感测到的手势信号,识别手势;以及确定与手势关联的控制指令。
其中,控制指令用于指示声音播放操作。声音播放操作是指声音播放设备执行的有关于声音播放的相关操作。例如,声音播放的相关操作可以是开启声音播放、关闭声音播放、音量调大、音量调小、播放上一个声音文件(例如歌曲文件)、或者播放下一个声音文件等。
控制装置100根据识别出的手势关联的控制指令,对声音播放设备进行控制,声音播放设备根据该控制指令执行相应的声音播放操作。声音播放设备可能是扬声器、或音乐播放器等。
该声音播放设备与控制装置100之间的可能的结构关系如下所述。
1)如图2所示,该控制装置100可以包括声音播放设备104。
可选的,声音播放设备104与处理器102连接,或者声音播放设备104、感应传感器101、处理器102或存储器103通过总线连接。
控制装置100可能是耳机,可以是有线耳机或无线耳机。声音播放设备104为耳机中的扬声器。例如耳机可以根据手势关联的控制指令,调整扬声器的播放音量大小。
当然,控制装置100也可能是其它类型的装置,例如手表、运动手环或眼镜等装置。
2)如图3a所示,该控制装置100与声音播放设备104耦合在一起。控制装置100中的部分或全部模块可以集成在声音播放设备104中。例如,处理器102和存储器103在声音播放装置中。
可能的一种应用场景为,感应传感器101在耳机中,声音播放设备104、处理器102和存储器103集成在手机中。耳机中的感应传感器101感测手势信号,将手势信号传输给手机中的处理器102识别手势,处理器102根据识别出的手势,确定该手势关联的控制指令,根据该控制指令控制声音播放设备104进行相关的声音播放操作。
3)如图3b所示,该控制装置100也可以与声音播放设备104连接。连接方式可以是有线或无线连接的方式,比如通过火线、通用串行总线(universal serial bus,USB)、蓝牙、或射频(radio frequency,RF)连接。
例如,声音播放设备104可以为手机、音乐播放器、个人计算机(personal computer,PC)或平板电脑等通信设备。控制装置100根据识别出来的手势关联的控制指令,控制音乐播放器或者手机进行声音播放操作。
存储器103用于存储一组程序,处理器102在根据手势信号识别手势时,可以调用存储器103中存储的程序来识别手势。
控制指令还可以用于控制电子设备上的元件或者用户界面(user interface,UI)元件等,例如,按钮、滚轮、滑动块。
下面对控制装置100对声音播放的控制方法的一些可选的实施方式进行详细描述。
感应传感器101可以感应到物体的靠近、远离、速度、距离以及各种手势。本申请实施例所述的手势信号的产生不限于通过人类的手做出的各类手势,也可以是其它物体的移动来实现的手势。例如,其它物体为笔,通过笔的靠近或远离来实现手势,也可以通过笔的转动来实现手势,比如笔的顺时针转动和逆时针转动实现不同的手势。
参考图4的场景来讲,以感应传感器101为基于多普勒效应的声波传感器为例,对控制装置和声音播放的控制方法做进一步详细说明。
感应传感器101(声波传感器)识别手势信号,手势信号可以为声波信号。感应传感器101也可以称为声波收发装置或声波收发器。感应传感器101发射声波,可以是周期性的发射声波。当感应传感器101的感应范围内出现手势时,该感应传感器101可以识别出手势给声波带来的变化,例如波形或幅度等变化。根据该变化可以识别出手势信号。也就是说手势信号是一种声波信号,包括声波的波形、或幅度等信息。
感应传感器101将该声波信号传递给处理器102。处理器102根据手势信号,识别手势。处理器102可以先对声波信号做以下至少一种处理:将接收到的声波信号转换为电信号;该电信号为模拟信号;模数转换(analogue-to-digital conversion,ADC),即将模拟信号转换为数字信号;对数字信号进行处理。
可选的,处理器102中可以包括信号处理模块和其它处理模块。由信号处理模块来实现上述对声波信号所做的处理过程。或者,如图4中虚线部分所示,处理器102与信号处理模块连接,由与处理器102连接的信号处理模块接收声波传感器的声波信号,将接收到的声波信号转换为电信号;该电信号为模拟信号;ADC,即将模拟信号转换为数字信号;对数字信号进行处理,并将处理后的信号再传递给处理器102,本申请这里不做限定。
处理器102可以根据处理后的手势信号识别手势。手势即用户在感应传感器101感应 范围内做出的各中手势。如图4中所示,处理器102在根据手势信号识别手势时,可以调用存储器103中存储的程序来识别手势。处理器102根据识别的手势,输出手势对应的控制指令。可选的,处理器102根据识别的手势对应的控制指令,控制计算机系统的不同的元件,例如扬声器。或者,处理器102根据识别的手势对应的控制指令,控制电子设备以及该电子设备的相关元件。在一些情况下,外接电子设备或外接手持通信设备,可能需要通过连接装置实现连接。连接装置可以是有线或者无线连接,比如火线,USB,蓝牙,RF。
本申请实施例中,手势与控制指令具有关联关系。本申请实施例中,可以预先存储多个手势与多个控制指令的关联关系,具体地,这个关联关系可以存储在控制装置100本地的存储介质中,当然也可以存储在控制装置100本地以外的其他远程存储介质中。
其中,一个手势可以对应一个控制指令。当然也可能多个手势对应一个控制指令,当控制装置感测到多个手势中的任意一个手势时,均可以实现该控制指令。也可能一个手势对应多个控制指令,当控制装置感测到该手势时,可以实现该手势对应的多个控制指令,例如,一个手势对应切换到下一个声音文件并调大音量,当控制装置感测到该手势时,确定切换到下一个声音文件并调大下一个声音文件的音量。
在本申请实施例中,存储器103可以是非易失性存储器,比如硬盘(hard disk drive,HDD)或固态硬盘(solid-state drive,SSD)等,还可以是易失性存储器(volatile memory),例如随机存取存储器(random-access memory,RAM)。存储器是能够用于携带或存储具有指令或数据结构形式的期望的程序代码并能够由计算机存取的任何其他介质,但不限于此。本申请实施例中的存储器还可以是电路或者其它任意能够实现存储功能的装置,用于存储程序指令和/或数据。可选的,手势与控制信令之间的关联关系可以存储在存储器103中。存储器103还可以存储用于手势关联控制指令的程序。存储器103存储的程序可以是一个单独的程序,也可以是操作系统的一部分。
在本申请实施例中,处理器102可以是通用处理器、数字信号处理器、专用集成电路、现场可编程门阵列或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件。通用处理器可以是微处理器或者任何常规的处理器等。
处理器102在确定与手势关联的控制指令后,该控制指令控制声音播放设备104执行相应的操作。声音播放设备104可能是计算机系统的元件,与控制装置100位于同一个设备中,例如控制装置100置于耳机中,声音播放设备104为耳机中的扬声器。或者,控制装置100连接到声音播放设备104,例如,控制装置置于耳机中,声音播放设备104为平板电脑中的相关元件或设备,通过耳机中的控制装置的控制指令,控制平板电脑中的相关元件或设备执行控制指令相关的操作。
声音播放设备104在执行控制指令所指示的操作的过程中,可以向用户提供反馈信息。该反馈信息可以体现在音频、触觉或显示等至少一个方面。例如,声音播放设备104根据控制指令将手机上音乐播放器的播放音量调大,在手机屏幕上可以显示音量调大的界面。
本申请实施例中,根据手势信号识别手势的方法可以采用任意的方法。示例性的,本申请实施例可以通过以下方式来实现识别手势。
用户执行手势姿态,感应传感器101发射声波,收到手势的声波信号。感应传感器101将声波信号传递给处理器102,处理器102将手势的声波信号转换为手势的电信号。将手势的模拟信号转化为手势的数字信号。手势的数字信号可以通过手势波形显示。根据手势的数字信号识别出手势。
处理器102将手势信号(可以是数字信号)转化为特征的集合或特征的列表。其中,每个特征代表一个输入,比如一个特征代表一个手势。一般来说,每个特征可以对应一些信息,包括以下任意一种或多种:标识(ID)、x轴时间、y轴微多普勒频率或z轴幅度。通过这些信息来表示一个特征。例如,第一特征由ID1,x1,y1,z1表示,第二特征由ID2,x2,y2,z2表示。从手势信号(如手势的数字信号)到特征的转换可以利用数字信号处理算法。
处理器102进行特征分类和分组。在分类过程中,确定每个特征的身份或类别,例如手指、手掌、手臂或者其它物体。一旦被分类,多个特征就可以被分到相应的特征组。一个特征组内的特征中的位置可能彼此邻近,或者一个特征组内的特征来自相同的手指。可选的,分组的过程还可包括过滤,通过过滤来滤除不是手势姿态事件的一部分特征。过滤可以提高手势识别的精确度。
处理器102确定特征组的关键参数,关键参数可以包括特征之间的距离、特征的质心、特征旋转、状态转变、特征的运动速率或特征的运动方向等,确定特征组的关键参数可由识别算法完成。
通过以上过程,处理器102能够识别出唯一手势。
可选的,如果手势为人类手势,处理器102可以通过以下方式来识别出人类手势。
提供与人类手势相关的数据的时间序列。将时间序列的数据转换成包括右手波形和左手波形的形式(波形)。从波形中提取特征,其中,提取步骤包括从包括手位置的波形中提取静态特征,从波形中提取动态特征,包括从右侧波形中提取相位和频率,从左侧波形中提取相位和频率。根据提取的特征,将右手波形中的一个或多个手的位置、相位和频率,左手波形中的手的位置、相位和频率,合成一条描述手势的规则来识别手势。由特征或特征的组合组成的规则,用来唯一地识别人类的手势。
上文基于控制装置100的模块采用硬件的形式实现来描述的。可选的,控制装置100的模块还可以采用软件的形式来实现。
示例性地,如图5所示,当控制装置的模块采用软件的形式来实现时,控制装置500包括识别模块501和确定模块502。
识别模块501用于根据手势信号,识别手势;
确定模块502,用于确定与所述手势关联的所述控制指令,所述控制指令用于指示声音播放操作。
可选的,在根据所述手势信号,识别手势时,识别模块501用于:将声波信号转换为电信号;根据所述电信号,识别手势。
在一个可能的设计中,识别模块501可以通过以下方式来识别手势:进行特征分类和分组。在分类过程中,确定每个特征的身份或类别,例如手指、手掌、手臂或者其它物体。一旦被分类,多个特征就可以被分到相应的特征组。一个特征组内的特征中的位置可能彼此邻近或者来自相同的手指。可选的,分组的过程可包括过滤,通过过滤来滤除不是手势姿态事件的一部分特征。过滤可以提高手势识别的精确度。识别模块501确定特征组的关键参数。关键参数包括特征之间的距离、特征的质心、特征旋转、状态转变、特征的运动速率或特征的运动方向等,确定特征组的关键参数可由识别算法完成。通过以上过程,识别模块501能够识别出唯一手势。
识别模块501和确定模块502执行的操作可以参照上述处理器102执行的相应操作。
该控制装置500还包括发送模块503,发送模块503用于将控制指令发送给声音播放设备;或者,该装置还包括执行模块504,所述执行模块504用于执行所述控制指令所指示的声音播放操作。发送模块可以通过有线或无线的方式发送控制指令。控制装置500与声音播放设备之间的架构关系也可以参考上述控制装置100与声音播放设备104之间的架构关系。
可选的,控制装置500还可能包括存储模块505,存储模块505的功能参考存储器103的功能,在此不再赘述。
控制装置还可能采用软件和硬件结合的形式来实现。例如,处理器中的硬件及软件模块组合执行完成处理器执行的上述操作。
通过本申请实施例提供的控制装置,使得对声音播放控制可以满足更多的应用场景。用户不需要触摸电子设备或声音播放设备或元件,用户也不需要语音输入,而是通过手势就可以产生控制指令来控制声音播放。应用场景更加宽泛。例如,对于一些语言障碍的人群,在不方便通过语音输入控制指令时,可以借用手势来输入控制指令。对于通过光感来输入控制指令,光感的环境限制因素很多,容易受到环境光信号干扰,而通过手势来输入控制指令,可以避免环境的干扰,更加方便且控制精准。相比通过物理触摸来输入控制指令,通过手势来输入控制指令的方法可以避免耳机尺寸很小操作不便的问题,操作更加方便。
以下给出一些手势与控制指令的关联的示例。当感应传感器101识别到以下举例的手势时,处理器102可以通过上述处理确定与手势对应的控制指令。可以理解的是,本申请实施例并不限于这些手势以及对应的控制指令。以下描述的各个手势可以任意多个结合使用。各个手势可以是左手形成的手势,也可以是右手形成的手势。如果控制装置为耳机,左手形成的手势可以用于控制左耳耳机,右手形成的手势可以控制右耳耳机,当然也可以任意手形成的手势控制任意耳机。
(1)手势为握拳,控制指令为开启播放。握拳是指手指向掌心弯曲成拳头。如图6所示,从a、b、c、d、到e表示握拳的手势姿态。
(2)手势为拳头张开,控制指令为关闭播放或停止播放。如图6所示,从e、d、c、b、到a表示右手拳头张开的手势姿态。
(3)手势为大拇指与其他手指靠近,控制指令为开启播放。其中,其他手指包括食指、中指、无名指或小拇指中的一个或多个。如图7所示,从a到b表示为右手大拇指与食指靠近的手势。
大拇指与其他手指靠近可以是,大拇指保持不动,其他手指向大拇指靠近;也可以是其他手指保持不动,大拇指向其他手指靠近;也可以是大拇指和其他手指相互靠近。
该手势也可以将大拇指替换为其他类型的任意一个手指。
(4)手势为大拇指与其他手势远离,控制指令为关闭或停止播放。其中,其他手指包括食指、中指、无名指或小拇指中的一个或多个。如图7所示,从b到a表示为右手大拇指与食指远离的手势。
大拇指与其他手指远离可以是,大拇指保持不动,其他手指向大拇指远离;也可以是其他手指保持不动,大拇指向其他手指远离;也可以是大拇指和其他手指相互远离。
该手势也可以将大拇指替换为其他类型的任意一个手指。
基于(3)和(4)的手势,手指靠近可以一直到捏合状态,也可以靠近到某一个距离 但并未捏合的状态。类似的,手指远离可以从捏合状态开始远离,也可以从某一个距离但并未捏合的状态开始远离。
在一个可能的实现方式中,(3)所描述的手势,也可以关联控制指令为音量调小。
在一个可能的实现方式中,(4)所描述的手势,也可以关联控制指令为音量调大。
例如,距离越大音量越大,距离越小音量越小。手指捏合对应静音状态。
(5)手势为手指向上运动,控制指令为音量增大。手指可以是任意一个或多个手指。如图8所示,以右手大拇指为例,按照箭头向上的方向运动表示手指向上运动的手势。
(6)手势为手指向下运动,控制指令为音量减小。手指可以是任意一个或多个手指。如图8所示,以右手大拇指为例,按照箭头向下的方向运动标识手指向下运动的手势。
(7)手势为手指向前运动,控制指令为音量增大。手指可以是任意一个或多个手指。如图9所示,以左手食指为例,从a到b表示手指向前运动的手势。
(8)手势为手指向后运动,控制指令为音量减小。手指可以是任意一个或多个手指。如图9所示,以左手食指为例,从b到a表示手指向后运动的手势。
(9)手势为手指向左运动,控制指令为音量增大。手指可以是任意一个或多个手指。如图10所示,以左手食指为例,从a到b表示手指向左运动的手势。
(10)手势为手指向右运动,控制指令为音量减小。手指可以是任意一个或多个手指。如图10所示,以左手食指为例,从b到a表示手指向右运动的手势。
(11)手势为手势模拟旋转物理旋钮顺时针转动,控制指令为音量增大。如图11所示,顺时针转动实现该手势。该手势在实现时假设存在一个物理旋钮,用手来选择该物理旋钮体现出该手势。当然该物理旋钮是不存在的。
(12)手势为手势模拟旋转物理旋钮逆时针转动,控制指令为音量减小。仍如图11所示,逆时针转动实现该手势。
(13)手势为小臂向远离身体方向挥动,控制指令为切换播放下一个声音文件。如图12所示,从a到b表示小臂向远离身体方向挥动的手势。
(14)手势为小臂向靠近身体方向挥动,控制指令为切换播放上一个声音文件。如图12所示,从b到a表示小臂向远离身体方向挥动的手势。
(15)如图13所示:
手势一为左手靠近或远离左耳,控制指令为左耳机音量调整。例如,左手靠近左耳对应左耳机音量增大,右手远离右耳对应右耳机音量减小。
手势二为右手靠近或远离右耳,控制指令为右耳机音量调整。
手势一和手势二可以同时进行,手势一和手势二的运动方向可以是相同的,也可以是不同的。例如左手靠近左耳,右手靠近左耳。又例如,例如左手靠近左耳,右手远离左耳。
通过手势一和手势二的同时执行,可以实现立体音效的控制指令。
上述手势仅是一些举例,可以不限于上述举例的手势来实现各种控制指令。例如,还可以:手指靠近控制装置的手势,控制指令为音量增大;手指远离控制装置的手势,控制指令为音量减小。
上述手势手指可以指向任意方向,可选的,在实现手势时手指可以指向控制装置。
综上,以上描述的各个手势与各个控制指令的关系可以相互交换,例如,手势1关联控制指令1,手势2关联控制指令2,本申请实施例还可以将手势1关联控制指令2,手势 2关联控制指令1。比如手指向上运动,控制指令为音量增大,手指向下运动,控制指令为音量减小。比如,手势为小臂向远离身体方向挥动,控制指令为切换播放上一个声音文件;手势为小臂向靠近身体方向挥动,控制指令为切换播放下一个声音文件。
以上描述的任意两个或多个手势可以结合使用,组成一套手势控制指令。
本申请上述方法实施例描述的控制装置100或控制装置500所执行的操作和功能中的部分或全部,可以用芯片或集成电路来完成。
为了实现上述控制装置100或控制装置500的功能,本申请实施例还提供一种芯片,包括处理器,用于支持该控制装置100或控制装置500实现上述功能。在一种可能的设计中,该芯片与存储器连接或者该芯片包括存储器,该存储器用于保存该通信装置必要的程序指令和数据。
本申请实施例提供了一种计算机可读存储介质,存储有计算机程序,该计算机程序包括用于执行控制装置100或控制装置500所执行的方法的指令。
本申请实施例提供了一种包含指令的计算机程序产品,当其在计算机上运行时,使得计算机执行控制装置100或控制装置500所执行的方法。
基于同一技术构思,如图14所示,本申请实施例还提供一种声音播放的控制方法,该方法可以由上述控制装置来执行。该方法的具体流程如下所述。
S1401、感测手势信号,比如由控制装置中的感应传感器感测手势信号,具体感知过程请参照上述装置实施例中的详细描述,这里不予重复详述。
S1402、根据该手势信号,识别手势,比如由控制装置中的处理器根据该手势信号,识别手势,具体识别过程请参照上述装置实施例中的详细描述,这里不予重复详述。
S1403、确定与识别出来的手势相关联的控制指令,该控制指令用于指示声音播放操作。比如由控制装置中的处理器确定与识别出来的手势相关联的控制指令。
可选的,控制装置还可以根据识别出的手势关联的控制指令,对声音播放设备进行控制,例如可以将控制指令发送给声音播放设备。该声音播放设备接收到该控制指令后执行相关的声音播放操作。例如控制装置为耳机,耳机将控制指令发送给手机播放器,手机播放器执行该控制指令。此外,控制装置自身还可以根据识别出的手势关联的控制指令执行相应的声音播放操作。例如,控制装置为耳机,耳机自身来执行该控制指令,比方耳机扬声器声音调大或调小。
具体方法的细节可以参照上文中对控制装置的描述细节,重复之处不再赘述。
本领域内的技术人员应明白,本申请的实施例可提供为方法、系统、或计算机程序产品。因此,本申请可采用完全硬件实施例、完全软件实施例、或结合软件和硬件方面的实施例的形式。而且,本申请可采用在一个或多个其中包含有计算机可用程序代码的计算机可用存储介质(包括但不限于磁盘存储器、CD-ROM、光学存储器等)上实施的计算机程序产品的形式。
本申请是参照根据本申请实施例的方法、设备(系统)、和计算机程序产品的流程图和/或方框图来描述的。应理解可由计算机程序指令实现流程图和/或方框图中的每一流程和/或方框、以及流程图和/或方框图中的流程和/或方框的结合。可提供这些计算机程序指令到通用计算机、专用计算机、嵌入式处理机或其他可编程数据处理设备的处理器以产生一个机器,使得通过计算机或其他可编程数据处理设备的处理器执行的指令产生用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的 装置。
这些计算机程序指令也可存储在能引导计算机或其他可编程数据处理设备以特定方式工作的计算机可读存储器中,使得存储在该计算机可读存储器中的指令产生包括指令装置的制造品,该指令装置实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能。
这些计算机程序指令也可装载到计算机或其他可编程数据处理设备上,使得在计算机或其他可编程设备上执行一系列操作步骤以产生计算机实现的处理,从而在计算机或其他可编程设备上执行的指令提供用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的步骤。
尽管已描述了本申请的优选实施例,但本领域内的技术人员一旦得知了基本创造性概念,则可对这些实施例作出另外的变更和修改。所以,所附权利要求意欲解释为包括优选实施例以及落入本申请范围的所有变更和修改。
显然,本领域的技术人员可以对本申请实施例进行各种改动和变型而不脱离本申请实施例的精神和范围。这样,倘若本申请实施例的这些修改和变型属于本申请权利要求及其等同技术的范围之内,则本申请也意图包含这些改动和变型在内。

Claims (16)

  1. 一种声音播放的控制方法,其特征在于,所述方法包括:
    感测手势信号;
    根据所述手势信号,识别手势;
    确定与所述手势关联的控制指令,所述控制指令用于指示声音播放操作。
  2. 如权利要求1所述的方法,其特征在于,所述手势信号包括声波信号。
  3. 如权利要求2所述的方法,其特征在于,所述感测手势信号,包括:基于超声波多普勒效应,感测手势信号。
  4. 如权利要求1~3任一项所述的方法,其特征在于,所述控制指令为开启播放,所述手势为以下任意一种:握拳,大拇指与其他手指靠近;和/或,
    所述控制指令为停止播放,所述手势为以下任意一种:拳头张开,大拇指与其他手指远离;
    其中,所述其他手指包括食指、中指、无名指或小拇指中的一个或多个。
  5. 如权利要求1~4任一项所述的方法,其特征在于,所述控制指令为音量增大,所述手势为以下任意一种:手指向上运动,手指向前运动,手指向左运动,或手势模拟旋转物理旋钮顺时针转动;和/或,
    所述控制指令为音量减小,所述手势为以下任意一种:手指向下运动,手指向后运动,手指向右运动,或手势模拟旋转物理旋钮逆时针转动;
    其中,所述手指为任意一个或多个手指。
  6. 如权利要求1~5任一项所述的方法,其特征在于,
    所述手势为小臂向远离身体方向挥动,所述控制指令为切换播放下一个声音文件;和/或,
    所述手势为小臂向靠近身体方向挥动,所述控制指令为切换播放上一个声音文件。
  7. 如权利要求1~6任一项所述的方法,其特征在于,
    所述手势为左手靠近或远离左耳,所述控制指令为左耳机音量调整;和/或,
    所述手势为右手靠近或远离右耳,所述控制指令为右耳机音量调整。
  8. 一种控制装置,其特征在于,包括:
    感应传感器,用于感测手势信号;
    处理器,用于根据所述手势信号,识别手势;以及用于确定与所述手势关联的所述控制指令,所述控制指令用于指示声音播放操作。
  9. 如权利要求8所述的装置,其特征在于,所述手势信号包括声波信号。
  10. 如权利要求9所述的装置,其特征在于,所述感应传感器,具体用于基于超声波多普勒效应,感测手势信号。
  11. 如权利要求8~10任一项所述的装置,其特征在于,所述处理器在所述手势为握拳或大拇指与其他手指靠近时,确定所述控制指令为开启播放;和/或,
    所述处理器在所述手势为拳头张开或大拇指与其他手指远离时,确定所述控制指令为停止播放;
    其中,所述其他手指包括食指、中指、无名指或小拇指中的一个或多个。
  12. 如权利要求8~11任一项所述的装置,其特征在于,所述处理器在所述手势为手 指向上运动、手指向前运动、手指向左运动或手势模拟旋转物理旋钮顺时针转动中的任意一种时,确定所述控制指令为音量增大;和/或,
    所述处理器在所述手势为手指向下运动、手指向后运动、手指向右运动或手势模拟旋转物理旋钮逆时针转动中的任意一种时,确定所述控制指令为音量减小;
    其中,所述手指为任意一个或多个手指。
  13. 如权利要求8~12任一项所述的装置,其特征在于,
    所述处理器在所述手势为小臂向远离身体方向挥动时,确定所述控制指令为切换播放下一个声音文件;和/或,
    所述处理器在所述手势为小臂向靠近身体方向挥动时,确定所述控制指令为切换播放上一个声音文件。
  14. 如权利要求8~13任一项所述的装置,其特征在于,
    所述处理器在所述手势为左手靠近或远离左耳时,确定所述控制指令为左耳机音量调整;和/或,
    所述处理器在所述手势为右手靠近或远离右耳,确定所述控制指令为右耳机音量调整。
  15. 一种计算机可读存储介质,其特征在于,所述计算机存储介质中存储有计算机可读指令,当所述计算机可读指令在控制装置上运行时,使得所述控制装置执行权利要求1~7任一项所述的方法。
  16. 一种芯片,其特征在于,所述芯片与存储器相连或者所述芯片包括所述存储器,用于读取并执行所述存储器中存储的软件程序,以实现如权利要求1~7任一项所述的方法。
PCT/CN2020/139782 2019-12-26 2020-12-27 一种声音播放的控制方法及装置 WO2021129848A1 (zh)

Priority Applications (1)

Application Number Priority Date Filing Date Title
EP20907319.6A EP4068052A4 (en) 2019-12-26 2020-12-27 CONTROL METHOD AND DEVICE FOR AUDIO REPRODUCTION

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201911368976.3 2019-12-26
CN201911368976.3A CN113050788A (zh) 2019-12-26 2019-12-26 一种声音播放的控制方法及装置

Publications (1)

Publication Number Publication Date
WO2021129848A1 true WO2021129848A1 (zh) 2021-07-01

Family

ID=76505566

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/139782 WO2021129848A1 (zh) 2019-12-26 2020-12-27 一种声音播放的控制方法及装置

Country Status (3)

Country Link
EP (1) EP4068052A4 (zh)
CN (1) CN113050788A (zh)
WO (1) WO2021129848A1 (zh)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113347526B (zh) * 2021-07-08 2022-11-22 歌尔科技有限公司 耳机的音效调节方法、装置及可读存储介质
CN113630689B (zh) * 2021-08-06 2022-09-16 珠海格力电器股份有限公司 音量控制方法、装置、设备及存储介质
CN113923553A (zh) * 2021-09-28 2022-01-11 深圳市芯中芯科技有限公司 一种基于雷达成像的蓝牙耳机人机交互方法及装置
CN114051172B (zh) * 2022-01-11 2024-03-22 杭州阿里云飞天信息技术有限公司 一种直播交互方法、装置、电子设备及计算机程序产品

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160274732A1 (en) * 2015-03-16 2016-09-22 Elliptic Laboratories As Touchless user interfaces for electronic devices
CN109298791A (zh) * 2018-10-24 2019-02-01 Oppo广东移动通信有限公司 终端控制方法、装置、存储介质及移动终端
CN109753144A (zh) * 2018-05-08 2019-05-14 北京字节跳动网络技术有限公司 一种实现信息交互的方法、设备及系统
CN110045819A (zh) * 2019-03-01 2019-07-23 华为技术有限公司 一种手势处理方法及设备

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8749485B2 (en) * 2011-12-20 2014-06-10 Microsoft Corporation User control gesture detection
EP3133474B1 (en) * 2015-08-19 2019-03-27 Nxp B.V. Gesture detector using ultrasound
CN106446801B (zh) * 2016-09-06 2020-01-07 清华大学 基于超声主动探测的微手势识别方法及系统
CN109597312B (zh) * 2018-11-26 2022-03-01 北京小米移动软件有限公司 音箱控制方法及装置

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160274732A1 (en) * 2015-03-16 2016-09-22 Elliptic Laboratories As Touchless user interfaces for electronic devices
CN109753144A (zh) * 2018-05-08 2019-05-14 北京字节跳动网络技术有限公司 一种实现信息交互的方法、设备及系统
CN109298791A (zh) * 2018-10-24 2019-02-01 Oppo广东移动通信有限公司 终端控制方法、装置、存储介质及移动终端
CN110045819A (zh) * 2019-03-01 2019-07-23 华为技术有限公司 一种手势处理方法及设备

Also Published As

Publication number Publication date
EP4068052A1 (en) 2022-10-05
EP4068052A4 (en) 2023-01-18
CN113050788A (zh) 2021-06-29

Similar Documents

Publication Publication Date Title
WO2021129848A1 (zh) 一种声音播放的控制方法及装置
US10936081B2 (en) Occluded gesture recognition
JP6275839B2 (ja) リモコン装置、情報処理方法およびシステム
US20160349845A1 (en) Gesture Detection Haptics and Virtual Tools
US9632683B2 (en) Methods, apparatuses and computer program products for manipulating characteristics of audio objects by using directional gestures
US11237640B2 (en) Wearable device enabling multi-finger gestures
CN105388995A (zh) 启用触觉的可变形表面的形状输入和输出的系统和方法
US10636417B2 (en) Method and apparatus for performing voice recognition on basis of device information
JP2017509181A (ja) ジェスチャ相互作用式の装着可能な空間オーディオシステム
WO2019105376A1 (zh) 手势识别方法、终端及存储介质
US11023067B2 (en) Input control using fingerprints
KR20140114913A (ko) 사용자 기기의 센서 운용 방법 및 장치
WO2020078319A1 (zh) 基于手势的操控方法及终端设备
KR20150065336A (ko) 전자 장치의 제스처 인식 방법, 장치 및 컴퓨터 판독 가능한 기록 매체
WO2022001791A1 (zh) 一种基于ppg信息的智能设备交互方法
Li et al. Enabling voice-accompanying hand-to-face gesture recognition with cross-device sensing
CN108920052B (zh) 页面显示控制方法及相关产品
EP2990914A1 (en) User interaction with an apparatus using a location sensor and microphone signal(s)
KR102216358B1 (ko) 제스쳐를 이용하는 단말 제어 방법
TWI684120B (zh) 鼠標控制方法以及鼠標
WO2018079301A1 (ja) 情報処理装置、方法及びプログラム
US20170199578A1 (en) Gesture control method for interacting with a mobile or wearable device
KR102197912B1 (ko) 제스처 인식에 따른 기능 실행 방법, 장치 및 기록매체
CN108008902A (zh) 一种虚拟键盘的调整方法、装置和一种播放方法、装置
KR20240016663A (ko) 로봇 장갑과 가상현실 기기를 포함하는 가상 악기 연주 장치

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20907319

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2020907319

Country of ref document: EP

Effective date: 20220627

NENP Non-entry into the national phase

Ref country code: DE