CN110111786B - Audio processing method, equipment and computer readable storage medium - Google Patents

Audio processing method, equipment and computer readable storage medium Download PDF

Info

Publication number
CN110111786B
CN110111786B CN201910360041.4A CN201910360041A CN110111786B CN 110111786 B CN110111786 B CN 110111786B CN 201910360041 A CN201910360041 A CN 201910360041A CN 110111786 B CN110111786 B CN 110111786B
Authority
CN
China
Prior art keywords
audio processing
wearable device
area
voice
state
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910360041.4A
Other languages
Chinese (zh)
Other versions
CN110111786A (en
Inventor
夏小聪
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tibet Canyu Lunzhu Technology Co ltd
Original Assignee
Tibet Canyu Lunzhu Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tibet Canyu Lunzhu Technology Co ltd filed Critical Tibet Canyu Lunzhu Technology Co ltd
Priority to CN201910360041.4A priority Critical patent/CN110111786B/en
Publication of CN110111786A publication Critical patent/CN110111786A/en
Application granted granted Critical
Publication of CN110111786B publication Critical patent/CN110111786B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination

Abstract

The application discloses an audio processing method, equipment and a computer readable storage medium, wherein the method comprises the following steps: determining a wearing state of the wearable device; then, in the wearing state, recognizing a holding gesture of the wearing equipment in the voice receiving and transmitting process; then, according to the holding gesture and the ear position area, determining an audio processing area for receiving and transmitting voice in the wearable device; finally, an audio processing assembly is arranged in the audio processing area, wherein the audio processing assembly comprises a vibrating diaphragm, a coil and magnetic steel. The humanized audio processing scheme is realized, so that the user can conveniently send and receive the voice when using the wearable device, the definition of sending and receiving the voice is improved, and the user experience is enhanced.

Description

Audio processing method, equipment and computer readable storage medium
Technical Field
The present application relates to the field of mobile communications, and in particular, to an audio processing method, apparatus, and computer readable storage medium.
Background
In the prior art, with the rapid development of intelligent terminal equipment, wearable equipment different from conventional smart phones, such as wearable equipment like a smart watch or a smart bracelet, appears. Because wearing formula equipment compares in traditional smart mobile phone, its particularities such as software, hardware environment, operation mode and operation environment, if the scheme of controlling of traditional smart mobile phone is transferred to wearing formula equipment, can bring inconvenience, user experience bad for user's operation.
Disclosure of Invention
In order to solve the above technical drawbacks in the prior art, the present invention proposes an audio processing method, which includes:
determining a wearing state of the wearable device;
in the wearing state, recognizing a holding gesture of the wearing equipment in the voice receiving and transmitting process;
determining an audio processing area for receiving and transmitting voice in the wearable device according to the holding gesture and the ear position area;
and setting an audio processing assembly in the audio processing area, wherein the audio processing assembly comprises a vibrating diaphragm, a coil and magnetic steel.
Optionally, the determining the wearing state of the wearable device includes:
acquiring a connection state of the buckling component of the wearable equipment;
if the buckling component is in the connection state, the wearable device is determined to be in the wearing state, and if the buckling component is in the disconnection state, the wearable device is determined to be in the disconnection state.
Optionally, the determining the wearing state of the wearable device further includes:
in the wearing state, determining the current wearing position according to the buckling position of the buckling assembly and the buckled pressure sensing value;
And in the separation state, determining a current holding area according to the screen triggering area and/or the edge pressure sensing area of the wearable equipment.
Optionally, in the wearing state, the identifying the holding gesture of the wearable device in the voice receiving and transmitting process includes:
acquiring the operation instruction of voice receiving and transmitting;
and recognizing a holding gesture of the wearable device in the current state according to the operation instruction, wherein the holding gesture comprises a first holding gesture in the wearing state and a second holding gesture in the disengaging state.
Optionally, in the wearing state, the recognizing the holding gesture of the wearable device in the voice receiving and transmitting process further includes:
acquiring motion parameters of the wearable equipment in the voice receiving and transmitting process;
and adjusting a first audio processing parameter of gesture voice transceiving under the first holding gesture according to the motion parameter, or adjusting a second audio processing parameter of gesture voice transceiving under the second holding gesture according to the motion parameter.
Optionally, the determining, in the wearable device, the audio processing area for receiving and transmitting the voice according to the holding gesture and the ear position area includes:
Recognizing an ear position area of a voice transceiver through an ultrasonic assembly and/or an image recognition assembly of the wearable device;
in the first grip gesture state, an area of the wearable device associated with the ear position area is determined as the audio processing area, or in the second grip gesture state, an area of the wearable device associated with the ear position area is determined as the audio processing area.
Optionally, an audio processing component is disposed in the audio processing area, where the audio processing component includes a diaphragm, a coil, and magnetic steel, and includes:
disposing one or more sets of the audio processing components within one or more of the audio processing regions;
the audio processing assembly at least comprises two groups of coils and a plurality of groups of magnetic steel.
Optionally, an audio processing component is disposed in the audio processing area, where the audio processing component includes a diaphragm, a coil, and magnetic steel, and further includes:
the vibrating diaphragm, the first coil, the second coil and the magnetic steel are sequentially arranged in the audio processing area;
when the audio processing assembly receives a sounding signal, triggering the second coil to generate a first acting force in a magnetic field generated by the magnetic steel through the sounding signal, and driving the vibrating diaphragm through the first acting force to generate an acoustic wave signal;
When the vibrating diaphragm of the audio processing component receives the sound wave signal, the first coil is driven to vibrate, electromotive force is obtained through vibration in the magnetic field, and the electromotive force is converted into a sound receiving electric signal.
The invention also proposes an audio processing device comprising:
a memory, a processor, and a computer program stored on the memory and executable on the processor;
the computer program implementing the steps of the method according to any of the preceding claims when executed by the processor.
The invention also proposes a computer readable storage medium having stored thereon an audio processing program which, when executed by a processor, implements the steps of the audio processing method as described in any of the above.
The wearable device has the beneficial effects that the wearing state of the wearable device is determined; then, in the wearing state, recognizing a holding gesture of the wearing equipment in the voice receiving and transmitting process; then, according to the holding gesture and the ear position area, determining an audio processing area for receiving and transmitting voice in the wearable device; finally, an audio processing assembly is arranged in the audio processing area, wherein the audio processing assembly comprises a vibrating diaphragm, a coil and magnetic steel. The humanized audio processing scheme is realized, so that the user can conveniently send and receive the voice when using the wearable device, the definition of sending and receiving the voice is improved, and the user experience is enhanced.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the application and together with the description, serve to explain the principles of the application.
In order to more clearly illustrate the embodiments of the application or the technical solutions of the prior art, the drawings which are used in the description of the embodiments or the prior art will be briefly described, and it will be obvious to a person skilled in the art that other drawings can be obtained from these drawings without inventive effort.
Fig. 1 is a schematic hardware structure of an implementation manner of a wearable device according to an embodiment of the present application;
fig. 2 is a hardware schematic diagram of an implementation manner of a wearable device according to an embodiment of the present application;
fig. 3 is a hardware schematic of an implementation manner of a wearable device according to an embodiment of the present application;
fig. 4 is a hardware schematic of an implementation manner of a wearable device according to an embodiment of the present application;
fig. 5 is a hardware schematic of an implementation manner of a wearable device according to an embodiment of the present application;
FIG. 6 is a flow chart of a first embodiment of the audio processing method of the present application;
FIG. 7 is a flow chart of a second embodiment of the audio processing method of the present application;
FIG. 8 is a flow chart of a third embodiment of the audio processing method of the present invention;
FIG. 9 is a flow chart of a fourth embodiment of the audio processing method of the present invention;
FIG. 10 is a flowchart of a fifth embodiment of the audio processing method of the present invention;
FIG. 11 is a flowchart of a sixth embodiment of an audio processing method of the present invention;
FIG. 12 is a flowchart of a seventh embodiment of an audio processing method of the present invention;
FIG. 13 is a flowchart of an eighth embodiment of an audio processing method of the present invention;
fig. 14 is a block diagram of an audio processing component corresponding to the audio processing method of the present invention.
Detailed Description
It should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the scope of the invention.
In the following description, suffixes such as "module", "component", or "unit" for representing elements are used only for facilitating the description of the present invention, and have no specific meaning per se. Thus, "module," "component," or "unit" may be used in combination.
The wearable device provided by the embodiment of the invention comprises a mobile terminal such as an intelligent bracelet, an intelligent watch and an intelligent mobile phone. With the continuous development of screen technology, mobile terminals such as smart phones and the like can also be used as wearable devices due to the appearance of screen forms such as flexible screens, folding screens and the like. The wearable device provided in the embodiment of the invention can comprise: RF (Radio Frequency) unit, wiFi module, audio output unit, A/V (audio/video) input unit, sensor, display unit, user input unit, interface unit, memory, processor, and power supply.
In the following description, a wearable device will be taken as an example, please refer to fig. 1, which is a schematic hardware structure of a wearable device implementing various embodiments of the present invention, where the wearable device 100 may include: an RF (Radio Frequency) unit 101, a WiFi module 102, an audio output unit 103, an a/V (audio/video) input unit 104, a sensor 105, a display unit 106, a user input unit 107, an interface unit 108, a memory 109, a processor 110, and a power supply 111. Those skilled in the art will appreciate that the wearable device structure shown in fig. 1 does not constitute a limitation of the wearable device, and that the wearable device may include more or fewer components than shown, or certain components in combination, or a different arrangement of components.
The following describes the various components of the wearable device in detail with reference to fig. 1:
the radio frequency unit 101 may be used to send and receive information or send signals in a call process, specifically, the radio frequency unit 101 may send uplink information to the base station, or may send downlink information sent by the base station to the processor 110 of the wearable device to process the downlink information, where the downlink information sent by the base station to the radio frequency unit 101 may be generated according to the uplink information sent by the radio frequency unit 101, or may be actively pushed to the radio frequency unit 101 after detecting that the information of the wearable device is updated, for example, after detecting that the geographic position where the wearable device is located changes, the base station may send a notification of the change of the geographic position to the radio frequency unit 101 of the wearable device, after receiving the notification of the message, the radio frequency unit 101 may send the notification of the message to the processor 110 of the wearable device to process, and the processor 110 of the wearable device may control the notification of the message to be displayed on the display panel 1061 of the wearable device; typically, the radio frequency unit 101 includes, but is not limited to, an antenna, at least one amplifier, a transceiver, a coupler, a low noise amplifier, a duplexer, and the like. In addition, the radio frequency unit 101 may also communicate with a network and other devices through wireless communication, which may specifically include: through wireless communication with a server in a network system, for example, the wearable device can download file resources from the server through wireless communication, for example, an application program can be downloaded from the server, after the wearable device finishes downloading a certain application program, if the file resources corresponding to the application program in the server are updated, the server can push a message notification of the resource update to the wearable device through wireless communication so as to remind a user to update the application program. The wireless communication may use any communication standard or protocol, including but not limited to GSM (Global System of Mobile communication, global System for Mobile communications), GPRS (General Packet Radio Service ), CDMA2000 (Code Division Multiple Access, CDMA 2000), WCDMA (Wideband Code Division Multiple Access ), TD-SCDMA (Time Division-Synchronous Code Division Multiple Access, time Division synchronous code Division multiple Access), FDD-LTE (Frequency Division Duplexing-Long Term Evolution, frequency Division Duplex Long term evolution), and TDD-LTE (Time Division Duplexing-Long Term Evolution, time Division Duplex Long term evolution), etc.
In one embodiment, the wearable device 100 may access an existing communication network by inserting a SIM card.
In another embodiment, the wearable device 100 may access an existing communication network by setting an esim card (Embedded-SIM), and by adopting the esim card, the internal space of the wearable device may be saved and the thickness may be reduced.
It will be appreciated that although fig. 1 shows a radio frequency unit 101, it will be appreciated that the radio frequency unit 101 is not an essential component of a wearable device and may be omitted entirely as required within the scope of not changing the essence of the invention. The wearable device 100 may implement communication connection with other devices or communication networks through the wifi module 102 alone, which is not limited by the embodiment of the present invention.
WiFi belongs to a short-distance wireless transmission technology, and the wearable device can help a user to send and receive emails, browse webpages, access streaming media and the like through the WiFi module 102, so that wireless broadband Internet access is provided for the user. Although fig. 1 shows a WiFi module 102, it is understood that it does not belong to the necessary constitution of the wearable device, and can be omitted entirely as required within the scope of not changing the essence of the invention.
The audio output unit 103 may convert audio data received by the radio frequency unit 101 or the WiFi module 102 or stored in the memory 109 into an audio signal and output as sound when the wearable device 100 is in a call signal reception mode, a talk mode, a recording mode, a voice recognition mode, a broadcast reception mode, or the like. Also, the audio output unit 103 may also provide audio output (e.g., call signal reception sound, message reception sound, etc.) related to a specific function performed by the wearable device 100. The audio output unit 103 may include a speaker, a buzzer, and the like.
The a/V input unit 104 is used to receive an audio or video signal. The a/V input unit 104 may include a graphics processor (Graphics Processing Unit, GPU) 1041 and a microphone 1042, the graphics processor 1041 processing image data of still pictures or video obtained by an image capturing device (e.g., a camera) in a video capturing mode or an image capturing mode. The processed image frames may be displayed on the display unit 106. The image frames processed by the graphics processor 1041 may be stored in the memory 109 (or other storage medium) or transmitted via the radio frequency unit 101 or the WiFi module 102. The microphone 1042 can receive sound (audio data) via the microphone 1042 in a phone call mode, a recording mode, a voice recognition mode, and the like, and can process such sound into audio data. The processed audio (voice) data may be converted into a format output that can be transmitted to the mobile communication base station via the radio frequency unit 101 in the case of a telephone call mode. The microphone 1042 may implement various types of noise cancellation (or suppression) algorithms to cancel (or suppress) noise or interference generated in the course of receiving and transmitting the audio signal.
In one embodiment, the wearable device 100 includes one or more cameras, and by opening the cameras, capturing of images, photographing, video recording and other functions can be achieved, and the positions of the cameras can be set as required.
The wearable device 100 further comprises at least one sensor 105, such as a light sensor, a motion sensor, and other sensors. Specifically, the light sensor includes an ambient light sensor and a proximity sensor, wherein the ambient light sensor can adjust the brightness of the display panel 1061 according to the brightness of ambient light, and the proximity sensor can turn off the display panel 1061 and/or the backlight when the wearable device 100 moves to the ear. As one type of motion sensor, the accelerometer sensor can detect the acceleration in all directions (typically three axes), and can detect the gravity and direction when stationary, and can be used for applications for recognizing the gesture of a mobile phone (such as horizontal-vertical screen switching, related games, magnetometer gesture calibration), vibration recognition related functions (such as pedometer, knocking), and the like.
In one embodiment, the wearable device 100 further comprises a proximity sensor, by employing the proximity sensor, the wearable device is able to achieve non-contact manipulation, providing more modes of operation.
In one embodiment, the wearable device 100 further comprises a heart rate sensor, which when worn, enables detection of heart rate by being in close proximity to the user.
In one embodiment, the wearable device 100 may further include a fingerprint sensor, by reading a fingerprint, security verification or the like can be achieved.
The display unit 106 is used to display information input by a user or information provided to the user. The display unit 106 may include a display panel 1061, and the display panel 1061 may be configured in the form of a liquid crystal display (Liquid Crystal Display, LCD), an Organic Light-Emitting Diode (OLED), or the like.
In one embodiment, the display panel 1061 employs a flexible display screen, and the wearable device employing the flexible display screen is capable of bending when worn, thereby fitting more. Optionally, the flexible display screen may be an OLED screen body and a graphene screen body, and in other embodiments, the flexible display screen may also be other display materials, which is not limited to this embodiment.
In one embodiment, the display panel 1061 of the wearable device may take a rectangular shape for ease of wrapping when worn. In other embodiments, other approaches may be taken as well.
The user input unit 107 may be used to receive input numeric or character information and to generate key signal inputs related to user settings and function control of the wearable device. In particular, the user input unit 107 may include a touch panel 1071 and other input devices 1072. The touch panel 1071, also referred to as a touch screen, may collect touch operations thereon or thereabout by a user (e.g., operations of the user on the touch panel 1071 or thereabout by using any suitable object or accessory such as a finger, a stylus, etc.) and drive the corresponding connection device according to a predetermined program. The touch panel 1071 may include two parts of a touch detection device and a touch controller. The touch detection device detects the touch azimuth of a user, detects a signal brought by touch operation and transmits the signal to the touch controller; the touch controller receives touch information from the touch detection device, converts it into touch point coordinates, and sends the touch point coordinates to the processor 110, and can receive and execute commands sent from the processor 110. Further, the touch panel 1071 may be implemented in various types such as resistive, capacitive, infrared, and surface acoustic wave. The user input unit 107 may include other input devices 1072 in addition to the touch panel 1071. In particular, other input devices 1072 may include, but are not limited to, one or more of a physical keyboard, function keys (e.g., volume control keys, switch keys, etc.), a trackball, mouse, joystick, etc., as specifically not limited herein.
In one embodiment, the sides of the wearable device 100 may be provided with one or more buttons. The button can realize a plurality of modes such as short pressing, long pressing, rotation and the like, thereby realizing a plurality of operation effects. The number of the buttons can be multiple, and different buttons can be combined for use, so that multiple operation functions are realized.
Further, the touch panel 1071 may overlay the display panel 1061, and when the touch panel 1071 detects a touch operation thereon or thereabout, the touch panel 1071 is transferred to the processor 110 to determine the type of touch event, and then the processor 110 provides a corresponding visual output on the display panel 1061 according to the type of touch event. Although in fig. 1, the touch panel 1071 and the display panel 1061 are two independent components for implementing the input and output functions of the wearable device, in some embodiments, the touch panel 1071 may be integrated with the display panel 1061 to implement the input and output functions of the wearable device, which is not limited herein. For example, when a message notification of a certain application is received through the rf unit 101, the processor 110 may control the message notification to be displayed in a certain preset area of the display panel 1061, where the preset area corresponds to a certain area of the touch panel 1071, and may control the message notification displayed in the corresponding area on the display panel 1061 by performing a touch operation on the certain area of the touch panel 1071.
The interface unit 108 serves as an interface through which at least one external device can be connected with the wearable apparatus 100. For example, the external devices may include a wired or wireless headset port, an external power (or battery charger) port, a wired or wireless data port, a memory card port, a port for connecting a device having an identification module, an audio input/output (I/O) port, a video I/O port, an earphone port, and the like. The interface unit 108 may be used to receive input (e.g., data information, power, etc.) from an external device and transmit the received input to one or more elements within the wearable apparatus 100 or may be used to transmit data between the wearable apparatus 100 and the external device.
In one embodiment, the interface unit 108 of the wearable device 100 adopts a contact structure, and is connected with other corresponding devices through the contact, so as to realize functions of charging, connection and the like. The contact can also be waterproof.
Memory 109 may be used to store software programs as well as various data. The memory 109 may mainly include a storage program area that may store an operating system, application programs required for at least one function (such as a sound playing function, an image playing function, etc.), and a storage data area; the storage data area may store data (such as audio data, phonebook, etc.) created according to the use of the handset, etc. In addition, memory 109 may include high-speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other volatile solid-state storage device.
The processor 110 is a control center of the wearable device, connects various parts of the entire wearable device with various interfaces and lines, performs various functions of the wearable device and processes data by running or executing software programs and/or modules stored in the memory 109, and invoking data stored in the memory 109, thereby performing overall monitoring of the wearable device. Processor 110 may include one or more processing units; preferably, the processor 110 may integrate an application processor that primarily handles operating systems, user interfaces, applications, etc., with a modem processor that primarily handles wireless communications. It will be appreciated that the modem processor described above may not be integrated into the processor 110.
The wearable device 100 may further include a power source 111 (such as a battery) for powering the various components, and preferably, the power source 111 may be logically connected to the processor 110 through a power management system, so as to perform functions of managing charging, discharging, and power consumption management through the power management system.
Although not shown in fig. 1, the wearable device 100 may further include a bluetooth module or the like, which is not described herein. The wearable device 100 can be connected with other terminal devices through bluetooth to realize communication and information interaction.
Fig. 2 to fig. 4 are schematic structural diagrams of a wearable device according to an embodiment of the present application. The wearable device comprises a flexible screen. When the wearable device is unfolded, the flexible screen is in a strip shape; when the wearable device is in a wearing state, the flexible screen is bent to be annular. Fig. 2 and 3 show schematic structural diagrams of the wearable device screen when unfolded, and fig. 4 shows schematic structural diagrams of the wearable device screen when bent.
Based on the above embodiments, it can be seen that if the device is a wristwatch, a bracelet, or a wearable device, the screen of the device may not cover the watchband area of the device, or may cover the watchband area of the device. The application proposes an alternative embodiment, in which the device may be a wristwatch, a bracelet or a wearable device, comprising a screen and a connection. The screen may be a flexible screen and the connection may be a wristband. Alternatively, the screen of the device or the display area of the screen may be partially or fully overlaid on the wristband of the device. Fig. 5 is a schematic hardware diagram of an implementation manner of a wearable device according to an embodiment of the present application, where a screen of the device extends to two sides, and a part of the screen is covered on a watchband of the device. In other embodiments, the screen of the device may also be entirely covered on the watchband of the device, which is not limited to the embodiment of the present application.
Example 1
Fig. 6 is a flowchart of a first embodiment of the audio processing method of the present invention. A method of audio processing, the method comprising:
s1, determining the wearing state of the wearable device;
s2, in the wearing state, recognizing a holding gesture of the wearable device in the voice receiving and transmitting process;
s3, determining an audio processing area for receiving and transmitting voice in the wearable device according to the holding gesture and the ear position area;
s4, setting an audio processing assembly in the audio processing area, wherein the audio processing assembly comprises a vibrating diaphragm, a coil and magnetic steel.
In this embodiment, first, a wearing state of the wearable device is determined; then, in the wearing state, recognizing a holding gesture of the wearing equipment in the voice receiving and transmitting process; then, according to the holding gesture and the ear position area, determining an audio processing area for receiving and transmitting voice in the wearable device; finally, an audio processing assembly is arranged in the audio processing area, wherein the audio processing assembly comprises a vibrating diaphragm, a coil and magnetic steel.
Optionally, in this embodiment, the wearable device has multiple fastening modes, for example, a watch-chain fastening, an adsorption fastening, and the like, and the fastening state of the wearable device is identified by arranging a sensing component at one end or two ends of the fastening, and then the wearing state is identified by the fastening state;
Optionally, in this embodiment, according to different usage requirements and different usage scenarios of the user of the wearable device, the optimal audio transceiving position is determined according to the current grip gesture, for example, the grip gesture used when the user answers the phone call is different from the grip gesture used when the user listens to the voice message, and for example, the grip gestures used when the left hand and the right hand of the user are respectively used to listen to the voice message are different;
optionally, in this embodiment, the audio processing area is set at one or more positions of the wearable device, and it can be understood that, due to different use requirements and different use scenarios of the user, there are different optimal audio transceiving positions, the audio processing area is set at one or more positions correspondingly;
optionally, in this embodiment, at least one set of audio processing components is disposed in one or more audio processing areas, where the audio processing components include a diaphragm, a coil, and magnetic steel.
The method has the beneficial effects that the wearing state of the wearable device is determined; then, in the wearing state, recognizing a holding gesture of the wearing equipment in the voice receiving and transmitting process; then, according to the holding gesture and the ear position area, determining an audio processing area for receiving and transmitting voice in the wearable device; finally, an audio processing assembly is arranged in the audio processing area, wherein the audio processing assembly comprises a vibrating diaphragm, a coil and magnetic steel. The humanized audio processing scheme is realized, so that the user can conveniently send and receive the voice when using the wearable device, the definition of sending and receiving the voice is improved, and the user experience is enhanced.
Example two
Fig. 7 is a flowchart of a second embodiment of the audio processing method according to the present invention, based on the above embodiment, the determining the wearing state of the wearable device includes:
s11, acquiring a connection state of the buckling component of the wearable device;
and S12, if the buckling component is in a connection state, determining that the wearable device is in a wearing state, and if the buckling component is in a disconnection state, determining that the wearable device is in a disconnection state.
In this embodiment, first, a connection state of the fastening component of the wearable device is obtained; then, if the buckling component is in a connection state, the wearable device is determined to be in a wearing state, and if the buckling component is in a disconnection state, the wearable device is determined to be in a disconnection state.
Optionally, a pressure sensing assembly or a contact sensing assembly is arranged at one end or two ends of the buckling, so that the buckling state of the wearable device is identified, and then the wearing state is identified through the buckling state;
optionally, a temperature sensing assembly is arranged at one end or two ends of the buckling part, so that the wearing state of the wearable device is identified according to the body surface temperature.
The beneficial effects of the embodiment are that the connection state of the buckling component of the wearable equipment is obtained; then, if the buckling component is in a connection state, the wearable device is determined to be in a wearing state, and if the buckling component is in a disconnection state, the wearable device is determined to be in a disconnection state. The voice processing scheme is more humanized, so that the user can send and receive the voice more conveniently when using the wearable device, the definition of the sent and received voice is improved, and the user experience is enhanced.
Example III
Fig. 8 is a flowchart of a third embodiment of the audio processing method according to the present invention, based on the above embodiment, the determining the wearing state of the wearable device further includes:
s13, determining the current wearing position according to the buckling position of the buckling assembly and the buckled pressure sensing value in the wearing state;
s14, in the separation state, determining a current holding area according to a screen triggering area and/or an edge pressure sensing area of the wearable device.
In this embodiment, first, in the wearing state, the current wearing position is determined according to the fastening position of the fastening component and the fastening pressure sensing value; and then, in the separation state, determining the current holding area according to the screen triggering area and/or the edge pressure sensing area of the wearable device.
Optionally, in the disengaged state, determining a current holding area according to one or more preset screen triggering areas of the wearable device;
optionally, in the disengaged state, a current holding area is determined according to the pressure sensing areas of the two side edges of the wearable device.
The embodiment has the beneficial effects that the current wearing position is determined according to the buckling position of the buckling component and the buckled pressure sensing value in the wearing state; and then, in the separation state, determining the current holding area according to the screen triggering area and/or the edge pressure sensing area of the wearable device. The voice processing scheme is more humanized, so that the user can send and receive the voice more conveniently when using the wearable device, the definition of the sent and received voice is improved, and the user experience is enhanced.
Example IV
Fig. 9 is a flowchart of a fourth embodiment of an audio processing method according to the present invention, based on the foregoing embodiment, where in the wearing state, the step of recognizing a holding gesture of the wearable device in a voice transceiving process includes:
s21, acquiring an operation instruction of voice transceiving;
s22, recognizing a holding gesture of the wearable device in a current state according to the operation instruction, wherein the holding gesture comprises a first holding gesture in a wearing state and a second holding gesture in a disengaging state.
In this embodiment, first, an operation instruction for transmitting and receiving the voice is obtained; and then, recognizing a holding gesture of the wearable device in the current state according to the operation instruction, wherein the holding gesture comprises a first holding gesture in the wearing state and a second holding gesture in the disengaging state.
Optionally, the operation instruction is used to identify a holding gesture of the wearable device in a current state, where the holding gesture includes a first holding gesture in a wearing state and a second holding gesture in a disengaging state, and more specifically, in a fastening state, the wearable device may also be in a disengaging state, and in this state, the wearable device still performs voice transceiving in a holding manner, so that the gesture in this state is attributed to the second holding gesture in this embodiment.
The beneficial effects of the embodiment are that the operation instruction of receiving and transmitting the voice is obtained; and then, recognizing a holding gesture of the wearable device in the current state according to the operation instruction, wherein the holding gesture comprises a first holding gesture in the wearing state and a second holding gesture in the disengaging state. The voice processing scheme is more humanized, so that the user can send and receive the voice more conveniently when using the wearable device, the definition of the sent and received voice is improved, and the user experience is enhanced.
Example five
Fig. 10 is a flowchart of a fifth embodiment of an audio processing method according to the present invention, based on the foregoing embodiment, where in the wearing state, a holding gesture of the wearable device in a voice transceiving process is recognized, and the method further includes:
s23, acquiring motion parameters of the wearable equipment in the voice receiving and transmitting process;
s24, adjusting a first audio processing parameter of gesture voice transceiving under the first holding gesture according to the motion parameter, or adjusting a second audio processing parameter of gesture voice transceiving under the second holding gesture according to the motion parameter.
In this embodiment, first, in the voice transceiving process, a motion parameter of the wearable device is obtained; and then, adjusting a first audio processing parameter of gesture voice transceiving under the first holding gesture according to the motion parameter, or adjusting a second audio processing parameter of gesture voice transceiving under the second holding gesture according to the motion parameter.
Optionally, in the voice transceiving process, acquiring a motion parameter of the wearable device, where the motion parameter includes a sensing parameter of a tri-axis gyroscope, an acceleration sensing parameter, and the like;
Optionally, the first audio processing parameter of the gesture voice transceiver under the first holding gesture is adjusted according to the motion parameter, or the second audio processing parameter of the gesture voice transceiver under the second holding gesture is adjusted according to the motion parameter, wherein the first audio processing parameter and the second audio processing parameter comprise a voice playing parameter and a voice collecting parameter during voice transceiver.
The method has the beneficial effects that the motion parameters of the wearable equipment are obtained in the voice receiving and transmitting process; and then, adjusting a first audio processing parameter of gesture voice transceiving under the first holding gesture according to the motion parameter, or adjusting a second audio processing parameter of gesture voice transceiving under the second holding gesture according to the motion parameter. The voice processing scheme is more humanized, so that the user can send and receive the voice more conveniently when using the wearable device, the definition of the sent and received voice is improved, and the user experience is enhanced.
Example six
Fig. 11 is a flowchart of a sixth embodiment of an audio processing method according to the present invention, based on the foregoing embodiment, the determining, in the wearable device, the audio processing area for receiving and transmitting voice according to the holding gesture and the ear position area includes:
S31, recognizing an ear position area of a voice transceiver through an ultrasonic assembly and/or an image recognition assembly of the wearable device;
s32, in the first holding gesture state, determining an area, associated with the ear position area, of the wearable device as the audio processing area, or in the second holding gesture state, determining an area, associated with the ear position area, of the wearable device as the audio processing area.
In this embodiment, first, an ear position area of a voice transceiver is identified by an ultrasonic component and/or an image identification component of the wearable device; then, in the first grip gesture state, an area of the wearable device associated with the ear position area is determined as the audio processing area, or in the second grip gesture state, an area of the wearable device associated with the ear position area is determined as the audio processing area.
Optionally, the ultrasonic assembly and/or the image recognition assembly of the wearable device is used for recognizing an ear position area of the voice transceiver, wherein the recognized ear position area comprises an auricle area and/or an ear canal area.
The method has the advantages that through the ultrasonic assembly and/or the image recognition assembly of the wearable device, the ear position area of a voice transceiver is recognized; then, in the first grip gesture state, an area of the wearable device associated with the ear position area is determined as the audio processing area, or in the second grip gesture state, an area of the wearable device associated with the ear position area is determined as the audio processing area. The voice processing scheme is more humanized, so that the user can send and receive the voice more conveniently when using the wearable device, the definition of the sent and received voice is improved, and the user experience is enhanced.
Example seven
Fig. 12 is a flowchart of a seventh embodiment of an audio processing method according to the present invention, based on the foregoing embodiment, the audio processing component is disposed in the audio processing area, where the audio processing component includes a diaphragm, a coil, and a magnetic steel, and includes:
s41, setting one or more groups of audio processing components in one or more audio processing areas;
s42, the audio processing assembly at least comprises two groups of coils and a plurality of groups of magnetic steel.
In this embodiment, first, one or more sets of the audio processing components are disposed in one or more of the audio processing regions; then, the audio processing assembly at least comprises two groups of coils and a plurality of groups of magnetic steel.
Optionally, one or more groups of the audio processing components are disposed in one or more of the audio processing regions, where one or more groups of the audio processing components in different regions share the coil and one or more audio components of the magnetic steel.
The beneficial effects of this embodiment are that by disposing one or more sets of the audio processing components in one or more of the audio processing regions; then, the audio processing assembly at least comprises two groups of coils and a plurality of groups of magnetic steel. The voice processing scheme is more humanized, so that the user can send and receive the voice more conveniently when using the wearable device, the definition of the sent and received voice is improved, and the user experience is enhanced.
Example eight
Fig. 13 is a flowchart of an eighth embodiment of an audio processing method according to the present invention, based on the foregoing embodiment, the audio processing component is disposed in the audio processing area, where the audio processing component includes a diaphragm, a coil, and magnetic steel, and further includes:
S43, arranging the vibrating diaphragm, the first coil, the second coil and the magnetic steel in the audio processing area in sequence;
s44, when the audio processing assembly receives a sounding signal, triggering the second coil to generate a first acting force in a magnetic field generated by the magnetic steel through the sounding signal, and driving the vibrating diaphragm through the first acting force to generate an acoustic wave signal;
and S45, when the vibrating diaphragm of the audio processing assembly receives the sound wave signal, driving the first coil to vibrate, obtaining electromotive force through vibration in the magnetic field, and converting the electromotive force into a sound-receiving electric signal.
In this embodiment, first, the diaphragm, the first coil, the second coil, and the magnetic steel are sequentially arranged in the audio processing area; then, when the audio processing component receives a sounding signal, triggering the second coil to generate a first acting force in a magnetic field generated by the magnetic steel through the sounding signal, and driving the vibrating diaphragm through the first acting force to generate an acoustic wave signal; finally, when the vibrating diaphragm of the audio processing component receives the sound wave signal, the first coil is driven to vibrate, electromotive force is obtained through vibration in the magnetic field, and the electromotive force is converted into a sound receiving electric signal.
Fig. 14 is a block diagram of an audio processing component corresponding to the audio processing method of the present invention. When the device receives an audio signal, the audio signal passes through the coil 2, and generates a force in the magnetic field of the magnetic steel, and the force is brought to the vibrating diaphragm to compress air and generate sound waves. Wherein, the sound principle of receiving of audio frequency: when the device receives sound waves, the force generated by the sound waves acts on the vibrating diaphragm to cause the vibrating diaphragm to vibrate and drive the coil 1 to vibrate correspondingly, the coil 1 moves in the magnetic steel to generate electromotive force, and the sound signals are converted into electric signals. The coil 2 and the coil 1 of the device are externally connected with a processor 5, the processor has signals of the coil 2, and the signals of the coil 1 are respectively processed.
It can be appreciated that the scheme is a sound producing and receiving integrated technology. Since sound can be air-borne or medium-borne, not only is sound spoken from the mouth absorbed. The scheme is to absorb the sound coming out of the auditory canal and transmit the sound to the other party. The device (earpiece/microphone) is placed inside the ear and not outside. Not only in software but also structurally completely isolating external noise. Therefore, the voice of the user can be clearly transferred in a noisy environment, and the voice recognition rate of the artificial intelligent product can be improved.
The beneficial effects of the embodiment are that the vibrating diaphragm, the first coil, the second coil and the magnetic steel are sequentially distributed in the audio processing area; then, when the audio processing component receives a sounding signal, triggering the second coil to generate a first acting force in a magnetic field generated by the magnetic steel through the sounding signal, and driving the vibrating diaphragm through the first acting force to generate an acoustic wave signal; finally, when the vibrating diaphragm of the audio processing component receives the sound wave signal, the first coil is driven to vibrate, electromotive force is obtained through vibration in the magnetic field, and the electromotive force is converted into a sound receiving electric signal. The voice processing scheme is more humanized, so that the user can send and receive the voice more conveniently when using the wearable device, the definition of the sent and received voice is improved, and the user experience is enhanced.
Example nine
The invention also proposes an audio processing device comprising:
a memory, a processor, and a computer program stored on the memory and executable on the processor;
the computer program implementing the steps of the method according to any of the preceding claims when executed by the processor.
Specifically, in this embodiment, first, a wearing state of the wearable device is determined; then, in the wearing state, recognizing a holding gesture of the wearing equipment in the voice receiving and transmitting process; then, according to the holding gesture and the ear position area, determining an audio processing area for receiving and transmitting voice in the wearable device; finally, an audio processing assembly is arranged in the audio processing area, wherein the audio processing assembly comprises a vibrating diaphragm, a coil and magnetic steel.
Optionally, in this embodiment, the wearable device has multiple fastening modes, for example, a watch-chain fastening, an adsorption fastening, and the like, and the fastening state of the wearable device is identified by arranging a sensing component at one end or two ends of the fastening, and then the wearing state is identified by the fastening state;
optionally, in this embodiment, according to different usage requirements and different usage scenarios of the user of the wearable device, the optimal audio transceiving position is determined according to the current grip gesture, for example, the grip gesture used when the user answers the phone call is different from the grip gesture used when the user listens to the voice message, and for example, the grip gestures used when the left hand and the right hand of the user are respectively used to listen to the voice message are different;
Optionally, in this embodiment, the audio processing area is set at one or more positions of the wearable device, and it can be understood that, due to different use requirements and different use scenarios of the user, there are different optimal audio transceiving positions, the audio processing area is set at one or more positions correspondingly;
optionally, in this embodiment, at least one set of audio processing components is disposed in one or more audio processing areas, where the audio processing components include a diaphragm, a coil, and magnetic steel.
In this embodiment, first, a connection state of the fastening component of the wearable device is obtained; then, if the buckling component is in a connection state, the wearable device is determined to be in a wearing state, and if the buckling component is in a disconnection state, the wearable device is determined to be in a disconnection state.
Optionally, a pressure sensing assembly or a contact sensing assembly is arranged at one end or two ends of the buckling, so that the buckling state of the wearable device is identified, and then the wearing state is identified through the buckling state;
optionally, a temperature sensing assembly is arranged at one end or two ends of the buckling part, so that the wearing state of the wearable device is identified according to the body surface temperature.
In this embodiment, first, in the wearing state, the current wearing position is determined according to the fastening position of the fastening component and the fastening pressure sensing value; and then, in the separation state, determining the current holding area according to the screen triggering area and/or the edge pressure sensing area of the wearable device.
Optionally, in the disengaged state, determining a current holding area according to one or more preset screen triggering areas of the wearable device;
optionally, in the disengaged state, a current holding area is determined according to the pressure sensing areas of the two side edges of the wearable device.
In this embodiment, first, an operation instruction for transmitting and receiving the voice is obtained; and then, recognizing a holding gesture of the wearable device in the current state according to the operation instruction, wherein the holding gesture comprises a first holding gesture in the wearing state and a second holding gesture in the disengaging state.
Optionally, the operation instruction is used to identify a holding gesture of the wearable device in a current state, where the holding gesture includes a first holding gesture in a wearing state and a second holding gesture in a disengaging state, and more specifically, in a fastening state, the wearable device may also be in a disengaging state, and in this state, the wearable device still performs voice transceiving in a holding manner, so that the gesture in this state is attributed to the second holding gesture in this embodiment.
In this embodiment, first, in the voice transceiving process, a motion parameter of the wearable device is obtained; and then, adjusting a first audio processing parameter of gesture voice transceiving under the first holding gesture according to the motion parameter, or adjusting a second audio processing parameter of gesture voice transceiving under the second holding gesture according to the motion parameter.
Optionally, in the voice transceiving process, acquiring a motion parameter of the wearable device, where the motion parameter includes a sensing parameter of a tri-axis gyroscope, an acceleration sensing parameter, and the like;
optionally, the first audio processing parameter of the gesture voice transceiver under the first holding gesture is adjusted according to the motion parameter, or the second audio processing parameter of the gesture voice transceiver under the second holding gesture is adjusted according to the motion parameter, wherein the first audio processing parameter and the second audio processing parameter comprise a voice playing parameter and a voice collecting parameter during voice transceiver.
In this embodiment, first, an ear position area of a voice transceiver is identified by an ultrasonic component and/or an image identification component of the wearable device; then, in the first grip gesture state, an area of the wearable device associated with the ear position area is determined as the audio processing area, or in the second grip gesture state, an area of the wearable device associated with the ear position area is determined as the audio processing area.
Optionally, the ultrasonic assembly and/or the image recognition assembly of the wearable device is used for recognizing an ear position area of the voice transceiver, wherein the recognized ear position area comprises an auricle area and/or an ear canal area.
In this embodiment, first, one or more sets of the audio processing components are disposed in one or more of the audio processing regions; then, the audio processing assembly at least comprises two groups of coils and a plurality of groups of magnetic steel.
Optionally, one or more groups of the audio processing components are disposed in one or more of the audio processing regions, where one or more groups of the audio processing components in different regions share the coil and one or more audio components of the magnetic steel.
In this embodiment, first, the diaphragm, the first coil, the second coil, and the magnetic steel are sequentially arranged in the audio processing area; then, when the audio processing component receives a sounding signal, triggering the second coil to generate a first acting force in a magnetic field generated by the magnetic steel through the sounding signal, and driving the vibrating diaphragm through the first acting force to generate an acoustic wave signal; finally, when the vibrating diaphragm of the audio processing component receives the sound wave signal, the first coil is driven to vibrate, electromotive force is obtained through vibration in the magnetic field, and the electromotive force is converted into a sound receiving electric signal.
It can be appreciated that the scheme is a sound producing and receiving integrated technology. Since sound can be air-borne or medium-borne, not only is sound spoken from the mouth absorbed. The scheme is to absorb the sound coming out of the auditory canal and transmit the sound to the other party. The device (earpiece/microphone) is placed inside the ear and not outside. Not only in software but also structurally completely isolating external noise. Therefore, the voice of the user can be clearly transferred in a noisy environment, and the voice recognition rate of the artificial intelligent product can be improved.
The method has the beneficial effects that the wearing state of the wearable device is determined; then, in the wearing state, recognizing a holding gesture of the wearing equipment in the voice receiving and transmitting process; then, according to the holding gesture and the ear position area, determining an audio processing area for receiving and transmitting voice in the wearable device; finally, an audio processing assembly is arranged in the audio processing area, wherein the audio processing assembly comprises a vibrating diaphragm, a coil and magnetic steel. The humanized audio processing scheme is realized, so that the user can conveniently send and receive the voice when using the wearable device, the definition of sending and receiving the voice is improved, and the user experience is enhanced.
The folding display method, the device and the computer readable storage medium are implemented by determining the wearing state of the wearable device; then, in the wearing state, recognizing a holding gesture of the wearing equipment in the voice receiving and transmitting process; then, according to the holding gesture and the ear position area, determining an audio processing area for receiving and transmitting voice in the wearable device; finally, an audio processing assembly is arranged in the audio processing area, wherein the audio processing assembly comprises a vibrating diaphragm, a coil and magnetic steel. The humanized audio processing scheme is realized, so that the user can conveniently send and receive the voice when using the wearable device, the definition of sending and receiving the voice is improved, and the user experience is enhanced.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising one … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element.
The foregoing embodiment numbers of the present invention are merely for the purpose of description, and do not represent the advantages or disadvantages of the embodiments.
From the above description of the embodiments, it will be clear to those skilled in the art that the above-described embodiment method may be implemented by means of software plus a necessary general hardware platform, but of course may also be implemented by means of hardware, but in many cases the former is a preferred embodiment. Based on such understanding, the technical solution of the present invention may be embodied essentially or in a part contributing to the prior art in the form of a software product stored in a storage medium (e.g. ROM/RAM, magnetic disk, optical disk) comprising instructions for causing a terminal (which may be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.) to perform the method according to the embodiments of the present invention.
The embodiments of the present invention have been described above with reference to the accompanying drawings, but the present invention is not limited to the above-described embodiments, which are merely illustrative and not restrictive, and many forms may be made by those having ordinary skill in the art without departing from the spirit of the present invention and the scope of the claims, which are to be protected by the present invention.

Claims (4)

1. A method of audio processing, the method comprising:
determining a wearing state of the wearable device;
in the wearing state, recognizing a holding gesture of the wearing equipment in the voice receiving and transmitting process;
determining an audio processing area for voice receiving and transmitting in the wearable equipment according to the holding gesture and an ear position area of a voice transceiver, wherein the ear position area is an auricle area and/or an auditory canal area;
enabling an audio processing assembly in the audio processing region, wherein the audio processing assembly comprises a vibrating diaphragm, a coil and magnetic steel;
determining a wearing state of the wearable device, comprising:
acquiring a connection state of the buckling component of the wearable equipment;
if the buckling component is in a connection state, determining that the wearable device is in a wearing state, and if the buckling component is in a disconnection state, determining that the wearable device is in a disconnection state;
in the wearing state, determining the current wearing position according to the buckling position of the buckling assembly and the buckled pressure sensing value;
in the separation state, determining a current holding area according to a screen triggering area and/or an edge pressure sensing area of the wearable equipment;
In the wearing state, recognizing a holding gesture of the wearing device in the voice receiving and transmitting process, including:
acquiring the operation instruction of voice receiving and transmitting;
identifying a holding gesture of the wearable device in a current state according to the operation instruction, wherein the holding gesture comprises a first gesture in a wearing state and a second gesture in a disengaging state;
acquiring motion parameters of the wearable equipment in the voice receiving and transmitting process;
adjusting a first audio processing parameter of the wearable device in the process of voice receiving and transmitting under the first gesture according to the motion parameter, or adjusting a second audio processing parameter of the wearable device in the process of voice receiving and transmitting under the second gesture according to the motion parameter;
the determining, in the wearable device, the audio processing area for receiving and transmitting voice according to the holding gesture and the ear position area includes:
recognizing an ear position area of a voice transceiver through an ultrasonic assembly and/or an image recognition assembly of the wearable device;
determining, in a state of the first gesture or the second gesture, an area of the wearable device associated with the ear position area as the audio processing area;
The enabling of the audio processing component within the audio processing region previously comprises:
disposing one or more sets of the audio processing components within one or more of the audio processing regions;
the audio processing assembly at least comprises two groups of coils and a plurality of groups of magnetic steel.
2. The audio processing method of claim 1, wherein the enabling of audio processing components within the audio processing region further comprises, before:
the vibrating diaphragm, the first coil, the second coil and the magnetic steel are sequentially arranged in the audio processing area;
when the audio processing assembly receives a sounding signal, triggering the second coil to generate a first acting force in a magnetic field generated by the magnetic steel through the sounding signal, and driving the vibrating diaphragm through the first acting force to generate an acoustic wave signal;
when the vibrating diaphragm of the audio processing component receives the sound wave signal, the first coil is driven to vibrate, electromotive force is obtained through vibration in the magnetic field, and the electromotive force is converted into a sound receiving electric signal.
3. An audio processing apparatus, the apparatus comprising:
a memory, a processor, and a computer program stored on the memory and executable on the processor;
The computer program implementing the steps of the method according to claim 1 or 2 when executed by the processor.
4. A computer-readable storage medium, characterized in that the computer-readable storage medium has stored thereon an audio processing program which, when executed by a processor, implements the steps of the audio processing method according to claim 1 or 2.
CN201910360041.4A 2019-04-30 2019-04-30 Audio processing method, equipment and computer readable storage medium Active CN110111786B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910360041.4A CN110111786B (en) 2019-04-30 2019-04-30 Audio processing method, equipment and computer readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910360041.4A CN110111786B (en) 2019-04-30 2019-04-30 Audio processing method, equipment and computer readable storage medium

Publications (2)

Publication Number Publication Date
CN110111786A CN110111786A (en) 2019-08-09
CN110111786B true CN110111786B (en) 2023-11-10

Family

ID=67487736

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910360041.4A Active CN110111786B (en) 2019-04-30 2019-04-30 Audio processing method, equipment and computer readable storage medium

Country Status (1)

Country Link
CN (1) CN110111786B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117255283B (en) * 2023-11-16 2024-04-05 深圳市雅乐电子有限公司 Earphone and control method thereof

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1324033A (en) * 2000-05-15 2001-11-28 张越苏 Wearable radio multimedia miniature computer with Chinese LINUX operation system
JP2005191766A (en) * 2003-12-25 2005-07-14 Seiko Instruments Inc Portable electronic device
CN103809866A (en) * 2012-11-13 2014-05-21 联想(北京)有限公司 Operation mode switching method and electronic equipment
CN107409154A (en) * 2015-12-31 2017-11-28 深圳市柔宇科技有限公司 Flexible wearable device
WO2018095044A1 (en) * 2016-11-22 2018-05-31 华为技术有限公司 Audio input/output method and terminal device
WO2019053996A1 (en) * 2017-09-13 2019-03-21 ソニー株式会社 Headphone device
CN109561221A (en) * 2018-12-26 2019-04-02 努比亚技术有限公司 A kind of call control method, equipment and computer readable storage medium

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150230022A1 (en) * 2014-02-07 2015-08-13 Samsung Electronics Co., Ltd. Wearable electronic system

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1324033A (en) * 2000-05-15 2001-11-28 张越苏 Wearable radio multimedia miniature computer with Chinese LINUX operation system
JP2005191766A (en) * 2003-12-25 2005-07-14 Seiko Instruments Inc Portable electronic device
CN103809866A (en) * 2012-11-13 2014-05-21 联想(北京)有限公司 Operation mode switching method and electronic equipment
CN107409154A (en) * 2015-12-31 2017-11-28 深圳市柔宇科技有限公司 Flexible wearable device
WO2018095044A1 (en) * 2016-11-22 2018-05-31 华为技术有限公司 Audio input/output method and terminal device
WO2019053996A1 (en) * 2017-09-13 2019-03-21 ソニー株式会社 Headphone device
CN109561221A (en) * 2018-12-26 2019-04-02 努比亚技术有限公司 A kind of call control method, equipment and computer readable storage medium

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
穿戴式多媒体系统;王玉;张重阳;郑世宝;薛明;邵长彬;於家健;;电视技术(第08期);全文 *

Also Published As

Publication number Publication date
CN110111786A (en) 2019-08-09

Similar Documents

Publication Publication Date Title
CN109982179B (en) Audio signal output method and device, wearable device and storage medium
CN109946986B (en) Household appliance control method, wearable device and computer readable storage medium
CN110350935B (en) Audio signal output control method, wearable device and readable storage medium
CN110099337B (en) Bone conduction audio output mode adjustment method, wearable device and storage medium
CN109819506B (en) Intelligent connection method, wearable device and computer readable storage medium
CN110213810B (en) Wearable device control method, wearable device and computer-readable storage medium
CN110098845B (en) Message processing method, wearable device, mobile terminal and readable storage medium
CN110072295B (en) Dual-channel communication method, device, first terminal and medium
CN110225282B (en) Video recording control method, device and computer readable storage medium
CN110139270B (en) Wearable device pairing method, wearable device and computer readable storage medium
CN110069200B (en) Wearable device input control method, wearable device and storage medium
CN110086808B (en) Account login control method and device and computer readable storage medium
CN110111786B (en) Audio processing method, equipment and computer readable storage medium
CN110620875A (en) Screenshot control method and device in video shooting process and computer-readable storage medium
CN110007761B (en) Wearable device, control method thereof and computer readable storage medium
CN110187770B (en) Wrist strap voice control interaction method and device and computer readable storage medium
CN110138965B (en) Wearable device, number formatting method and device thereof, and storage medium
CN110069102B (en) Display area regulation and control method and device and computer readable storage medium
CN109976613B (en) Text input implementation method, wearable device and computer-readable storage medium
CN110169625B (en) Wearable equipment and intelligent wrist-watch
CN109982210B (en) Wearable device audio output method and device, wearable device and storage medium
CN110083205B (en) Page switching method, wearable device and computer-readable storage medium
CN110109605B (en) Image sharing method and device and computer readable storage medium
CN110187950B (en) Method for adjusting picture display position, wearable device and storage medium
CN110012164B (en) Sound playing method of equipment, wearable equipment and computer readable storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
TA01 Transfer of patent application right
TA01 Transfer of patent application right

Effective date of registration: 20230922

Address after: No. 3605, 5th Floor, Building 3, Micro World, No. 23 Century Avenue, Liuwu New District, Lhasa, Tibet Autonomous Region, 850000

Applicant after: Tibet Canyu Lunzhu Technology Co.,Ltd.

Address before: 518000 6-8 / F, 10-11 / F, 6 / F, 6-10 / F, C zone, Han's innovation building, No. 9018, Beihuan Avenue, high tech Zone, Nanshan District, Shenzhen City, Guangdong Province

Applicant before: NUBIA TECHNOLOGY Co.,Ltd.

GR01 Patent grant
GR01 Patent grant