CN115118812A - Information processing method, intelligent terminal and storage medium - Google Patents

Information processing method, intelligent terminal and storage medium Download PDF

Info

Publication number
CN115118812A
CN115118812A CN202210655732.9A CN202210655732A CN115118812A CN 115118812 A CN115118812 A CN 115118812A CN 202210655732 A CN202210655732 A CN 202210655732A CN 115118812 A CN115118812 A CN 115118812A
Authority
CN
China
Prior art keywords
audio
information
target
target audio
preset
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210655732.9A
Other languages
Chinese (zh)
Inventor
郭勇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Transsion Holdings Co Ltd
Original Assignee
Shenzhen Transsion Holdings Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Transsion Holdings Co Ltd filed Critical Shenzhen Transsion Holdings Co Ltd
Priority to CN202210655732.9A priority Critical patent/CN115118812A/en
Publication of CN115118812A publication Critical patent/CN115118812A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/72442User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality for playing music files
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/0486Drag-and-drop
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/165Management of the audio stream, e.g. setting of volume, audio stream path
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/451Execution arrangements for user interfaces
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R3/00Circuits for transducers, loudspeakers or microphones

Abstract

The application provides an information processing method, an intelligent terminal and a storage medium, wherein the method comprises the following steps: acquiring data of a target audio to obtain a reference environment audio; and the target audio and the reference environment audio meet preset conditions, the target audio is played in a mute mode, and preset information is output. By adopting the method provided by the application, the audio is converted into the preset information to be displayed under the condition that the condition of mute output audio is determined to be met, so that the method can be suitable for a noisy environment, the user is ensured to obtain the content of the audio, the display effect is improved, and the user experience is improved.

Description

Information processing method, intelligent terminal and storage medium
Technical Field
The present application relates to the field of information processing technologies, and in particular, to an information processing method, an intelligent terminal, and a storage medium.
Background
Along with the popularization of intelligent terminals (such as mobile phones and tablet computers), the functions of the intelligent terminals are more and more powerful, and more applications or services are installed in the intelligent terminals, so that great convenience is brought to daily life of people.
In the course of conceiving and implementing the present application, the inventors found that at least the following problems existed: people can listen to music and watch videos through the intelligent terminal. When people listen to music and watch videos, the sound played outside the intelligent terminal can be heard. When people listen to the sound played by the loudspeaker of the intelligent terminal in different noise environments, the volume played by the intelligent terminal can be manually adjusted due to unclear listening, and the user experience is poor.
The foregoing description is provided for general background information and is not admitted to be prior art.
Disclosure of Invention
In view of the foregoing technical problems, the present application provides an information processing method, an intelligent terminal, and a storage medium, where a currently played audio is silently played and preset information determined according to the audio is displayed when it is determined that a condition for silently outputting the audio is satisfied. On one hand, the method can lead the user to know the content of the played audio, simplify the operation of the user and improve the user experience. On the other hand, can provide convenience for the user in noisy environment, can be applicable to too noisy environment, the suitability is strong.
In order to solve the above technical problem, in a first aspect, the present application provides an information processing method, which is applicable to an intelligent terminal, and includes the following steps:
s10, acquiring data of the target audio to obtain a reference environment audio;
and S20, when the target audio frequency and the reference environment audio frequency meet the preset conditions, playing the target audio frequency in a mute mode, and outputting preset information.
Optionally, the step of S20 includes:
and the target audio and the reference environment audio meet the condition of mute audio output, the target audio is played in a mute mode, and display information is output according to a preset rule.
Optionally, the display information comprises at least one of:
first prompt information for prompting audio closing;
second prompting information used for prompting the audio-to-text information;
and text information and/or images corresponding to the target audio.
Optionally, the outputting the display information according to a preset rule includes:
determining or generating the target audio as the audio of the target application;
identifying whether text information and/or images corresponding to the target audio exist in an interface for playing the target audio to obtain an identification result;
and outputting the display information according to the type of the target application and/or the identification result.
Optionally, the outputting the display information according to the type of the target application and the recognition result includes at least one of:
outputting the first prompt information and/or the second prompt information when the target application is of a first type;
under the condition that the target audio is of a second type and the identification result is negative, outputting text information and/or images corresponding to the first prompt information and the target audio;
and outputting the first prompt information when the target audio is the audio of the second type application and the identification result is yes.
Optionally, before the step S20, the method further includes:
separating first environment information from the reference environment audio according to the target audio to obtain a first environment audio;
determining or generating first analog loudness information according to the target audio, and determining or generating second analog loudness information according to the first environmental audio;
and when the first analog loudness information and/or the second analog loudness information meet preset environmental conditions, determining or generating conditions meeting the mute output audio.
Optionally, the determining or generating first analog loudness information according to the target audio and determining or generating second analog loudness information according to the first ambient audio include:
acquiring a first audio parameter of the target audio and a second audio parameter of the first environment audio;
determining or generating the first analog loudness information of the first audio parameter according to a first corresponding relationship between analog loudness information and audio parameters;
and determining or generating the second analog loudness information of the second audio parameter according to the second corresponding relationship between the analog loudness information and the audio parameter.
Optionally, after the step S20, the method further includes:
acquiring second environment information to obtain second environment audio;
determining or generating third analog loudness information according to the second environmental audio;
when the first analog loudness information and/or the third analog loudness value meet a preset audio condition, determining or generating a condition meeting an audible playing condition;
and outputting the sound of the target audio.
Optionally, the output mode of the first prompt message and/or the second prompt message includes at least one of the following:
displaying the first prompt message and/or the second prompt message in a status bar;
displaying the first prompt message and/or the second prompt message in a first preset floating window;
displaying the first prompt message and/or the second prompt message in a preset area;
periodically displaying the first prompt message and/or the second prompt message with a first preset time length;
and displaying the first prompt message and/or the second prompt message in the preset area, wherein the duration of the first prompt message and/or the second prompt message is a second preset duration.
Optionally, the outputting text information and/or images corresponding to the target audio, where the target audio corresponds to a target image frame, includes:
acquiring text information corresponding to the target audio and a preset pattern of the image;
and outputting the text information and/or the image corresponding to the target audio in the preset mode under the condition of outputting the target image frame.
Optionally, the step S10 includes:
receiving the starting operation of the mute audio output function;
responding to the starting operation, recording the played target audio, wherein the playing volume of the target audio is not zero, and obtaining the reference environment audio.
Optionally, when the target audio and the reference environmental audio satisfy a preset condition, the target audio is played in a mute manner, and after preset information is output, the method further includes:
receiving an adjusting operation for adjusting text information and/or images corresponding to the target audio;
and adjusting the display mode of the text information and/or the image corresponding to the target audio to obtain updated text information and/or image.
Optionally, the adjusting operation includes at least one of:
an adjustment operation for a display position of the text information and/or the image;
adjusting the display style of the text information and/or the image;
triggering operation aiming at the text information and/or the adjusting control carried by the image;
closing display operation aiming at the text information and/or the text information in the image;
and adjusting the display mode of the text information and/or the image to be displayed in a second preset floating window.
In a second aspect, an embodiment of the present application provides an information processing apparatus, including:
the acquisition unit is used for acquiring data of the target audio to obtain a reference environment audio;
and the playing unit is used for determining that the target audio and the reference environment audio meet preset conditions, playing the target audio in a mute manner and outputting preset information.
In addition, in this aspect, reference may be made to the related matters of the first aspect in other alternative embodiments of the information processing apparatus, and details are not described here.
The application also provides an intelligent terminal, including: the information processing method comprises a memory and a processor, wherein the memory stores an information processing program, and the information processing program realizes the steps of any one of the information processing methods when being executed by the processor.
The present application also provides a computer-readable storage medium, which stores a computer program that, when executed by a processor, implements the steps of any of the information processing methods described above.
As described above, the information processing method of the present application, which is applicable to an intelligent terminal, includes the steps of: recording and acquiring data of a target audio to obtain a reference environment audio; determining that the target audio and the reference environmental audio satisfy a preset condition; and playing the target audio in a mute manner, and outputting preset information. Through the technical scheme, the audio can be silently played under the condition that the detected audio meets the requirement of the audio output in a silent mode, the content of the audio is displayed, and the problem that a user cannot hear playing sound in a noisy environment by adjusting the playing volume is solved. The operation of the user is simplified, the user can acquire the content of the played audio in a noisy environment, the adjusting effect is improved, and the user experience is improved.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the present application and together with the description, serve to explain the principles of the application. In order to more clearly illustrate the technical solutions of the embodiments of the present application, the drawings needed to be used in the description of the embodiments will be briefly described below, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without inventive exercise.
Fig. 1 is a schematic diagram of a hardware structure of an intelligent terminal implementing various embodiments of the present application;
fig. 2 is a communication network system architecture diagram according to an embodiment of the present application;
fig. 3 is a flowchart illustrating an information processing method according to the first embodiment;
fig. 4a is a timing diagram illustrating an information processing method according to the first embodiment;
fig. 4b is a user interface diagram of the information processing method shown according to the first embodiment;
fig. 4c is a schematic view of a user interface of the information processing method shown according to the first embodiment;
fig. 4d is a schematic view of a user interface of the information processing method shown according to the first embodiment;
fig. 4e is a user interface diagram of the information processing method shown according to the first embodiment;
fig. 5 is a flowchart illustrating an information processing method according to a second embodiment;
FIG. 6a is a schematic view of a user interface illustrating an adjustment operation according to a second embodiment;
FIG. 6b is a schematic view of a user interface illustrating an adjustment operation according to the second embodiment;
FIG. 6c is a schematic view of a user interface illustrating an adjustment operation according to the second embodiment;
FIG. 6d is a schematic view of a user interface illustrating an adjustment operation according to the second embodiment;
FIG. 6e is a schematic view of a user interface illustrating an adjustment operation according to the second embodiment;
FIG. 6f is a schematic view of a user interface illustrating an adjustment operation according to the second embodiment;
FIG. 6g is a schematic view of a user interface illustrating an adjustment operation according to the second embodiment;
fig. 7 is a flowchart illustrating an information processing method according to a third embodiment;
FIG. 8a is a schematic diagram of a user interface showing the mute output audio function being turned on according to a third embodiment;
FIG. 8b is a schematic diagram of a user interface showing the mute output audio function being turned on according to the third embodiment;
FIG. 8c is a schematic diagram of a user interface showing the mute output audio function being turned on according to the third embodiment;
FIG. 8d is a schematic diagram of a user interface showing the mute output audio function being turned on according to the third embodiment;
fig. 9 is a schematic structural diagram of an information processing apparatus according to an embodiment of the present application.
The implementation, functional features and advantages of the objectives of the present application will be further explained with reference to the accompanying drawings. Specific embodiments of the present application have been shown by way of example in the drawings and will be described in more detail below. The drawings and written description are not intended to limit the scope of the inventive concepts in any manner, but rather to illustrate the concepts of the application by those skilled in the art with reference to specific embodiments.
Detailed Description
Reference will now be made in detail to the exemplary embodiments, examples of which are illustrated in the accompanying drawings. When the following description refers to the accompanying drawings, like numbers in different drawings represent the same or similar elements unless otherwise indicated. The embodiments described in the following exemplary embodiments do not represent all embodiments consistent with the present application. Rather, they are merely examples of apparatus and methods consistent with certain aspects of the present application, as detailed in the appended claims.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, the recitation of an element by the phrase "comprising an … …" does not exclude the presence of additional like elements in the process, method, article, or apparatus that comprises the element, and further, where similarly-named elements, features, or elements in different embodiments of the disclosure may have the same meaning, or may have different meanings, that particular meaning should be determined by their interpretation in the embodiment or further by context with the embodiment.
It should be understood that although the terms first, second, third, etc. may be used herein to describe various information, such information should not be limited to these terms. These terms are only used to distinguish one type of information from another. For example, first information may also be referred to as second information, and similarly, second information may also be referred to as first information, without departing from the scope herein. The word "if" as used herein may be interpreted as "at … …" or "when … …" or "in response to a determination", depending on the context. Also, as used herein, the singular forms "a," "an," and "the" are intended to include the plural forms as well, unless the context indicates otherwise. It will be further understood that the terms "comprises," "comprising," "includes" and/or "including," when used in this specification, specify the presence of stated features, steps, operations, elements, components, items, species, and/or groups, but do not preclude the presence, or addition of one or more other features, steps, operations, elements, components, species, and/or groups thereof. The terms "or," "and/or," "including at least one of the following," and the like, as used herein, are to be construed as inclusive or mean any one or any combination. For example, "includes at least one of: A. b, C "means" any of the following: a; b; c; a and B; a and C; b and C; a and B and C ", again for example," A, B or C "or" A, B and/or C "means" any of the following: a; b; c; a and B; a and C; b and C; a and B and C'. An exception to this definition will occur only when a combination of elements, functions, steps or operations are inherently mutually exclusive in some way.
It should be understood that, although the steps in the flowcharts in the embodiments of the present application are shown in order as indicated by the arrows, the steps are not necessarily performed in order as indicated by the arrows. The steps are not performed in the exact order shown and may be performed in other orders unless explicitly stated herein. Moreover, at least some of the steps in the figures may include multiple sub-steps or multiple stages that are not necessarily performed at the same time, but may be performed at different times, in different orders, and may be performed alternately or at least partially with respect to other steps or sub-steps of other steps.
The words "if", as used herein, may be interpreted as "at … …" or "at … …" or "in response to a determination" or "in response to a detection", depending on the context. Similarly, the phrases "if determined" or "if detected (a stated condition or event)" may be interpreted as "when determined" or "in response to a determination" or "when detected (a stated condition or event)" or "in response to a detection (a stated condition or event)", depending on the context.
It should be noted that, step numbers such as 301 and 302 are used herein for the purpose of more clearly and briefly describing the corresponding content, and do not constitute a substantial limitation on the sequence, and those skilled in the art may perform 302 first and then 301 in specific implementation, but these should be within the scope of the present application.
It should be understood that the specific embodiments described herein are merely illustrative of and not restrictive on the broad application.
In the following description, suffixes such as "module", "component", or "unit" used to denote elements are used only for the convenience of description of the present application, and have no specific meaning in themselves. Thus, "module", "component" or "unit" may be used mixedly.
The smart terminal may be implemented in various forms. For example, the smart terminal described in the present application may include smart terminals such as a mobile phone, a tablet computer, a notebook computer, a palmtop computer, a Personal Digital Assistant (PDA), a Portable Media Player (PMP), a navigation device, a wearable device, a smart band, a pedometer, and the like, and fixed terminals such as a Digital TV, a desktop computer, and the like.
While the following description will be given by way of example of a smart terminal, those skilled in the art will appreciate that the configuration according to the embodiments of the present application can be applied to a fixed type terminal in addition to elements particularly used for mobile purposes.
Referring to fig. 1, which is a schematic diagram of a hardware structure of an intelligent terminal for implementing various embodiments of the present application, the intelligent terminal 100 may include: RF (Radio Frequency) unit 101, WiFi module 102, audio output unit 103, a/V (audio/video) input unit 104, sensor 105, display unit 106, user input unit 107, interface unit 108, memory 109, processor 110, and power supply 111. Those skilled in the art will appreciate that the intelligent terminal architecture shown in fig. 1 does not constitute a limitation of the intelligent terminal, and that the intelligent terminal may include more or fewer components than shown, or some components may be combined, or a different arrangement of components.
The following specifically introduces each component of the intelligent terminal with reference to fig. 1:
the radio frequency unit 101 may be configured to receive and transmit signals during information transmission and reception or during a call, and specifically, receive downlink information of a base station and then process the downlink information to the processor 110; in addition, the uplink data is transmitted to the base station. Typically, radio frequency unit 101 includes, but is not limited to, an antenna, at least one amplifier, a transceiver, a coupler, a low noise amplifier, a duplexer, and the like. In addition, the radio frequency unit 101 can also communicate with a network and other devices through wireless communication. The wireless communication may use any communication standard or protocol, including but not limited to GSM (Global System for Mobile communications), GPRS (General Packet Radio Service), CDMA2000(Code Division Multiple Access 2000), WCDMA (Wideband Code Division Multiple Access), TD-SCDMA (Time Division-Synchronous Code Division Multiple Access), FDD-LTE (Frequency Division duplex-Long Term Evolution), TDD-LTE (Time Division duplex-Long Term Evolution, Time Division Long Term Evolution), 5G, and so on.
WiFi belongs to short-distance wireless transmission technology, and the intelligent terminal can help a user to receive and send e-mails, browse webpages, access streaming media and the like through the WiFi module 102, and provides wireless broadband internet access for the user. Although fig. 1 shows the WiFi module 102, it is understood that it does not belong to the essential constitution of the smart terminal, and may be omitted entirely as needed within the scope not changing the essence of the invention.
The audio output unit 103 may convert audio data received by the radio frequency unit 101 or the WiFi module 102 or stored in the memory 109 into an audio signal and output as sound when the smart terminal 100 is in a call signal reception mode, a call mode, a recording mode, a voice recognition mode, a broadcast reception mode, or the like. Also, the audio output unit 103 may also provide audio output related to a specific function performed by the smart terminal 100 (e.g., a call signal reception sound, a message reception sound, etc.). The audio output unit 103 may include a speaker, a buzzer, and the like.
The a/V input unit 104 is used to receive audio or video signals. The a/V input Unit 104 may include a Graphics Processing Unit (GPU) 1041 and a microphone 1042, the Graphics processor 1041 Processing image data of still pictures or video obtained by an image capturing device (e.g., a camera) in a video capturing mode or an image capturing mode. The processed image frames may be displayed on the display unit 106. The image frames processed by the graphic processor 1041 may be stored in the memory 109 (or other storage medium) or transmitted via the radio frequency unit 101 or the WiFi module 102. The microphone 1042 may receive sounds (audio data) via the microphone 1042 in a phone call mode, a recording mode, a voice recognition mode, or the like, and may be capable of processing such sounds into audio data. The processed audio (voice) data may be converted into a format output transmittable to a mobile communication base station via the radio frequency unit 101 in case of a phone call mode. The microphone 1042 may implement various types of noise cancellation (or suppression) algorithms to cancel (or suppress) noise or interference generated in the course of receiving and transmitting audio signals.
The smart terminal 100 also includes at least one sensor 105, such as a light sensor, a motion sensor, and other sensors. Optionally, the light sensor includes an ambient light sensor and a proximity sensor, the ambient light sensor may adjust the brightness of the display panel 1061 according to the brightness of ambient light, and the proximity sensor may turn off the display panel 1061 and/or the backlight when the smart terminal 100 moves to the ear. As one of the motion sensors, the accelerometer sensor can detect the magnitude of acceleration in each direction (generally, three axes), can detect the magnitude and direction of gravity when stationary, and can be used for applications of recognizing the posture of a mobile phone (such as horizontal and vertical screen switching, related games, magnetometer posture calibration), vibration recognition related functions (such as pedometer and tapping), and the like; as for other sensors such as a fingerprint sensor, a pressure sensor, an iris sensor, a molecular sensor, a gyroscope, a barometer, a hygrometer, a thermometer, and an infrared sensor, which can be configured on the mobile phone, further description is omitted here.
The display unit 106 is used to display information input by a user or information provided to the user. The Display unit 106 may include a Display panel 1061, and the Display panel 1061 may be configured in the form of a Liquid Crystal Display (LCD), an Organic Light-Emitting Diode (OLED), or the like.
The user input unit 107 may be used to receive input numeric or character information and generate key signal inputs related to user settings and function control of the intelligent terminal. Alternatively, the user input unit 107 may include a touch panel 1071 and other input devices 1072. The touch panel 1071, also referred to as a touch screen, may collect a touch operation performed by a user on or near the touch panel 1071 (e.g., an operation performed by the user on or near the touch panel 1071 using a finger, a stylus, or any other suitable object or accessory), and drive a corresponding connection device according to a predetermined program. The touch panel 1071 may include two parts of a touch detection device and a touch controller. Optionally, the touch detection device detects a touch orientation of a user, detects a signal caused by a touch operation, and transmits the signal to the touch controller; the touch controller receives touch information from the touch sensing device, converts the touch information into touch point coordinates, sends the touch point coordinates to the processor 110, and can receive and execute commands sent by the processor 110. In addition, the touch panel 1071 may be implemented in various types, such as a resistive type, a capacitive type, an infrared ray, and a surface acoustic wave. In addition to the touch panel 1071, the user input unit 107 may include other input devices 1072. Optionally, other input devices 1072 may include, but are not limited to, one or more of a physical keyboard, function keys (e.g., volume control keys, switch keys, etc.), a trackball, a mouse, a joystick, and the like, and are not limited thereto.
Alternatively, the touch panel 1071 may cover the display panel 1061, and when the touch panel 1071 detects a touch operation thereon or nearby, the touch panel 1071 transmits the touch operation to the processor 110 to determine the type of the touch event, and then the processor 110 provides a corresponding visual output on the display panel 1061 according to the type of the touch event. Although the touch panel 1071 and the display panel 1061 are shown in fig. 1 as two separate components to implement the input and output functions of the smart terminal, in some embodiments, the touch panel 1071 and the display panel 1061 may be integrated to implement the input and output functions of the smart terminal, and is not limited herein.
The interface unit 108 serves as an interface through which at least one external device is connected to the smart terminal 100. For example, the external device may include a wired or wireless headset port, an external power supply (or battery charger) port, a wired or wireless data port, a memory card port, a port for connecting a device having an identification module, an audio input/output (I/O) port, a video I/O port, an earphone port, and the like. The interface unit 108 may be used to receive input (e.g., data information, power, etc.) from an external device and transmit the received input to one or more elements within the smart terminal 100 or may be used to transmit data between the smart terminal 100 and the external device.
The memory 109 may be used to store software programs as well as various data. The memory 109 may mainly include a program storage area and a data storage area, and optionally, the program storage area may store an operating system, an application program required for at least one function (such as a sound playing function, an image playing function, and the like), and the like; the storage data area may store data (such as audio data, a phonebook, etc.) created according to the use of the cellular phone, and the like. Further, the memory 109 may include high speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other volatile solid state storage device.
The processor 110 is a control center of the intelligent terminal, connects various parts of the entire intelligent terminal using various interfaces and lines, and performs various functions of the intelligent terminal and processes data by operating or executing software programs and/or modules stored in the memory 109 and calling data stored in the memory 109, thereby performing overall monitoring of the intelligent terminal. Processor 110 may include one or more processing units; preferably, the processor 110 may integrate an application processor and a modem processor, optionally, the application processor mainly handles operating systems, user interfaces, application programs, etc., and the modem processor mainly handles wireless communications. It will be appreciated that the modem processor described above may not be integrated into the processor 110.
The intelligent terminal 100 may further include a power supply 111 (such as a battery) for supplying power to each component, and preferably, the power supply 111 may be logically connected to the processor 110 through a power management system, so as to implement functions of managing charging, discharging, and power consumption through the power management system.
Although not shown in fig. 1, the smart terminal 100 may further include a bluetooth module or the like, which is not described herein.
In order to facilitate understanding of the embodiments of the present application, a communication network system on which the intelligent terminal of the present application is based is described below.
Referring to fig. 2, fig. 2 is an architecture diagram of a communication Network system provided in an embodiment of the present application, where the communication Network system is an LTE system of a universal mobile telecommunications technology, and the LTE system includes a UE (User Equipment) 201, an E-UTRAN (Evolved UMTS Terrestrial Radio Access Network) 202, an EPC (Evolved Packet Core) 203, and an IP service 204 of an operator, which are in communication connection in sequence.
Optionally, the UE201 may be the intelligent terminal 100, which is not described herein again.
The E-UTRAN202 includes eNodeB2021 and other eNodeBs 2022, among others. Alternatively, the eNodeB2021 may be connected with other enodebs 2022 through a backhaul (e.g., X2 interface), the eNodeB2021 is connected to the EPC203, and the eNodeB2021 may provide the UE201 access to the EPC 203.
The EPC203 may include an MME (Mobility Management Entity) 2031, an HSS (Home Subscriber Server) 2032, other MMEs 2033, an SGW (Serving gateway) 2034, a PGW (PDN gateway) 2035, and a PCRF (Policy and Charging Rules Function) 2036, and the like. Optionally, the MME2031 is a control node that handles signaling between the UE201 and the EPC203, providing bearer and connection management. HSS2032 is used to provide registers to manage functions such as home location register (not shown) and holds subscriber specific information about service characteristics, data rates, etc. All user data may be sent through SGW2034, PGW2035 may provide IP address allocation and other functions for UE201, PCRF2036 is a policy and charging control policy decision point for traffic data flow and IP bearer resources, which selects and provides available policy and charging control decisions for policy and charging enforcement function (not shown).
The IP services 204 may include the internet, intranets, IMS (IP Multimedia Subsystem), or other IP services, among others.
Although the LTE system is described as an example, it should be understood by those skilled in the art that the present application is not limited to the LTE system, but may also be applied to other wireless communication systems, such as GSM, CDMA2000, WCDMA, TD-SCDMA, and future new network systems (e.g. 5G), and the like.
Based on the above intelligent terminal hardware structure and communication network system, various embodiments of the present application are provided.
The information processing method, the intelligent terminal, and the storage medium provided in the embodiments of the present application are further described in detail below. Referring to fig. 3, fig. 3 is a flowchart illustrating an information processing method according to a first embodiment. The information processing method shown in FIG. 3 includes steps 301 to 302. The method of the embodiment of the present application may be executed by the intelligent terminal shown in fig. 1, or may be executed by a chip in the intelligent terminal, and the intelligent terminal may be applied to the communication network system shown in fig. 2. The method execution subject shown in fig. 3 is an example of an intelligent terminal. Wherein:
301. and acquiring data of the target audio to obtain the reference environment audio.
In the embodiment of the application, the intelligent terminal can play videos and videos, and when the videos and the audios are played, the intelligent terminal can play audio and video sounds through a built-in loudspeaker of the intelligent terminal, so that a user can listen to music or watch videos through the intelligent terminal. When a user plays audio or video in a noisy environment, the played content may not be heard clearly because of the low volume. In some implementations, a distance sensor built in the intelligent terminal can detect a distance between the intelligent terminal and a user, and adjust the volume of playing according to the distance, or the intelligent terminal can acquire the noise intensity in the current environment, and automatically increase the volume of playing when the environmental noise is high, or otherwise decrease the volume of playing. However, the noise of the user in the noisy environment is too large, and after the volume is adjusted, the user may still not hear or hear the played sound clearly, the adjustment effect is not good, and the user experience is poor.
Optionally, the target audio may be any one of an audio played by the intelligent terminal or a video played by the intelligent terminal, which is not limited in this application. The reference environmental audio may be an audio obtained by recording a target audio played by the intelligent terminal. The intelligent terminal can acquire the data of the target audio to obtain the reference environment audio under the condition of playing the target audio. Optionally, the data of the target audio acquired by the intelligent terminal may be sound obtained by recording and playing the target audio, so as to obtain a reference environmental audio, and the intelligent terminal may determine whether a preset condition is met according to the target audio and the reference environmental audio, where the preset condition may be a condition for muting the output audio. Under the condition that the preset condition (namely the condition of mute audio output) is determined to be met, the intelligent terminal can play the target audio in a mute mode and output the preset information. Optionally, the preset information may be prompt information, or text information and/or images determined according to the target audio (text information and/or images corresponding to the target audio), so as to achieve the purpose of silent play. Optionally, the reference environmental audio recorded by the intelligent terminal includes both the target audio that is externally played by the intelligent terminal through a speaker and the environmental sound. The ambient sound may also be understood as ambient noise.
Alternatively, the mute playing refers to playing the content of the target audio, but not playing the sound, and displaying the currently played information (such as text information and/or image) of the target audio in the user interface, so as to ensure that the user knows the content of the target audio. Optionally, before the target audio is played, the intelligent terminal may receive an opening operation of a mute audio output function, and in response to the opening operation, record the played target audio to obtain the reference environmental audio under the condition that the target audio is played, and optionally, the playing volume of the target audio is not zero. Optionally, after the mute audio output function is turned on, the smart terminal may record the played audio under the condition that the audio is played each time, so as to determine whether the condition for mute audio output is satisfied. If the preset information is not satisfied, the sound of the audio is played normally.
302. And when the target audio and the reference environment audio meet preset conditions, playing the target audio in a mute manner, and outputting preset information.
Optionally, the reference environmental audio recorded by the intelligent terminal includes data of two audio frequencies, one is data of environmental noise of an environment where the intelligent terminal is currently located, and the other is data of sound of a target audio currently played by the intelligent terminal. The intelligent terminal can determine whether the condition of mute output audio is met according to the target audio and the reference environmental audio, and play the target audio in a mute manner and output preset information under the condition that the preset condition (namely the condition of mute output audio) is met according to the target audio and the reference environmental audio. Conversely, when it is determined that the condition for outputting audio in a mute state is satisfied according to the target audio and the reference environmental audio, that is, when the condition for playing audio in a voiced state is satisfied, the sound of the target audio is output, that is, the volume for playing the target audio at this time is not zero.
Optionally, the intelligent terminal determines whether a condition for muting the output audio is satisfied according to the target audio and the reference environmental audio, and the intelligent terminal may specifically separate current environmental noise, that is, first environmental information, from the reference environmental audio according to the target audio, so as to obtain the first environmental audio. Optionally, the first environment information may be separated by a machine learning method in the intelligent terminal, that is, the intelligent terminal may preset an audio separation model, and the audio separation model may be used to identify and separate various environment sounds to obtain separated first environment information, that is, first environment audio. Optionally, the intelligent terminal may also separate the target audio from the first environmental audio by using the difference in frequency characteristics, the intelligent terminal may preset or construct at least two band pass filters according to the target audio, the sound of the target audio only can pass through the first band pass filter, the first environmental audio only can pass through the second band pass filter, and then the separated first environmental information is used as the first environmental audio.
Optionally, the intelligent terminal may determine or generate first analog loudness information according to the target audio, determine or generate second analog loudness information according to the first environmental audio, and determine that the condition for muting the output audio is satisfied when the first analog loudness information and/or the second analog loudness information satisfy a preset environmental condition. The intelligent terminal can obtain a first audio parameter of the target audio and a second audio parameter of the first environment audio, determine first analog loudness information corresponding to the first audio parameter according to a first corresponding relation between the analog loudness information and the audio parameter, and determine second model loudness information corresponding to the second audio parameter according to a second corresponding relation between the analog loudness information and the audio parameter.
Alternatively, the audio parameters may include frequency characteristics and amplitude characteristics of the audio, the first audio parameter may be frequency characteristics of the target audio separated in the reference environmental audio, and the second audio parameter may be frequency characteristics of the first environmental audio separated in the reference environmental audio. The purpose of obtaining the analog loudness information is to perform mixed simulation in an actual scene, so as to simulate whether the user hears the sound played by the intelligent terminal clearly. Optionally, the first corresponding relationship and the second corresponding relationship may be preset in the intelligent terminal, or may be obtained, which is not limited in this application.
Optionally, the first corresponding relationship may represent an effect of playing the target audio with different first audio parameters, which is that the effect played by the intelligent terminal is different after the software algorithm and the intelligent terminal hardware gain are added during the playing process of the intelligent terminal, so that the effect heard by the actual user can be simulated through the first corresponding relationship. It will be appreciated that the first audio parameter of the recorded target audio is different when the target audio is played out at different volumes.
Optionally, the second corresponding relationship may also represent a noise degree of an actual environmental sound of the first environmental audio of the environment where the intelligent terminal is located under the influence of the intelligent terminal hardware device and the software algorithm, that is, the environmental noise of the environment where the actual user is located is simulated. Because the frequencies and amplitudes in the first audio parameter and the second audio parameter may not be in the same standard, the first audio parameter and the second audio parameter cannot be compared, and therefore the first audio parameter and the second audio parameter can be put in the same standard to be compared according to the first corresponding relation and the second corresponding relation, and the first analog loudness information and the second analog loudness information are obtained. Alternatively, the first analog loudness information and the second analog loudness information may be represented by a single value or a vector including at least one value, so that whether the preset environmental condition is satisfied or not may be determined according to the first analog loudness information and/or the second analog loudness information, and when the preset environmental condition is satisfied, the condition that the output audio is muted, that is, the preset condition is determined or generated.
Optionally, taking the presentation form of the first analog loudness information and the presentation form of the second analog loudness information as an example, when the first analog loudness information is greater than the second analog loudness information, it is simulated that when a user listens to the target audio in the environment, the sound of the target audio is dominant, that is, the preset audio condition is met, that is, the voiced playing condition is met, the preset environment condition is not met, and the condition of playing the audio in a mute manner is not met, then the target audio can be played at a normal volume, that is, the sound of the target audio is output. And/or under the condition that the first analog loudness information is smaller than the second analog loudness information, when the user listens to the target audio in the environment, the sound of the environmental audio is superior, namely the preset environmental condition is met, the condition of playing the audio in a mute manner is met, the preset audio condition is not met, namely the sound playing condition is not met, the target audio can be played in a mute manner, and the preset information is output.
Optionally, the preset information output by the intelligent terminal may be display information output according to a preset rule. Optionally, the display information may include at least one of: the audio-to-text prompting method comprises first prompting information used for prompting audio closing, second prompting information used for prompting audio to be converted into text information, and text information and/or images corresponding to target audio. Optionally, the first prompt information may be used to prompt only for a mute output in a subtitle scene; the second prompt message can be used for prompting that the mute output is needed in a subtitle-free scene; the text information and/or image corresponding to the target audio may be information that needs to be output in the subtitle-free scene.
Optionally, when a preset condition is met, that is, a condition for muting an output audio is met, the sound of the target audio may be muted, and since directly muting the target audio is more abrupt, so that user experience is affected, an interval duration may exist between playing the sound of the target audio and silently playing the target audio, that is, a duration for playing the target audio is a set duration, and a value of the set duration may be 3 seconds, 5 seconds, and the like, which is not limited in the present application. Before the target audio is output in a mute manner, the intelligent terminal may further output prompt information for prompting the user that the target audio is to be output in a mute manner, such as the first prompt information and/or the second prompt information.
Optionally, after the target audio is played for the set duration, the volume of the currently played target audio is not zero, and the target audio is output in a mute manner. Furthermore, the intelligent terminal may determine whether the preset condition (i.e., the condition for playing the audio in a mute manner) is met or whether the preset environment condition (i.e., the condition for playing the audio in a mute manner) is met again, if the preset condition (i.e., the condition for playing the audio in a mute manner) is still met, continue to mute and output the target audio, and if it is determined that the condition for playing the audio is met, play the sound of the target audio, where the played volume is not zero. Alternatively, the sound may be played back from the current playing position of the target audio, or may be played back from the playing position at which the target audio was played in the last mute mode, and played back.
Optionally, before the intelligent terminal plays the sound of the target audio, the intelligent terminal may further output a prompt message for prompting the user that the target audio is to be played audibly, and may continue to play the target audio in a mute manner, where a duration of the mute is a set duration. Alternatively, the set time period also serves to prevent the sound from being suddenly played, and the user feels less good, and thus, the set time period may be an interval time. Optionally, the set duration may be a fixed value, or may not be a fixed value, which is not limited in this application.
Referring to fig. 4a together, fig. 4a is a timing diagram of an information processing method according to a first embodiment, as shown in fig. 4a, the intelligent terminal first starts a mute audio output function, acquires data of a target audio, that is, a target audio that can be recorded and played by the intelligent terminal, obtains a reference environmental audio, determines whether a preset condition is satisfied according to the reference environmental audio and the target audio, and outputs a sound of the target audio, that is, performs normal voiced playing, when the target audio is not satisfied. If so, outputting prompt information (such as the first prompt information and/or the second prompt information), after continuing playing the sound of the target audio within a set time length, muting and outputting the target audio, and outputting preset information. Optionally, the user may also adjust the preset information. Optionally, the intelligent terminal may continuously determine whether a preset audio condition (i.e., a sound playing condition) is currently met, and if not, continue to play the audio in a mute manner, i.e., continue to play the target audio in a mute manner. If so, outputting prompt information to prompt the user that the sound is about to be played, continuing to play the mute sound for a set duration, and then playing the sound of the target audio for audio playing.
Optionally, the prompt information output by the intelligent terminal may include a first prompt information, a second prompt information, and a prompt information for prompting the user to play the audio immediately, where a manner of any one or more of the prompt information includes, but is not limited to, at least one of the following: the display method comprises the steps of displaying in a status bar, displaying in a first preset floating window, displaying in a preset area, periodically displaying in a first preset time length, and displaying in the preset area, wherein the displayed duration time length is a second preset time length. Optionally, the first preset duration and the second preset duration may be preset in the intelligent terminal, may be a fixed duration, or may not be a variable duration, which is not limited in the present application.
Referring to fig. 4b to 4c together, fig. 4b to 4c are schematic user interfaces illustrating the information processing method according to the first embodiment. For convenience of description, the application explains by taking the target audio corresponding to the target image frame as an example, that is, the target audio is the audio included in the target video as an example, as shown in the left side of fig. 4b, the intelligent terminal may display a playing interface of the target video, and display prompt information (such as first prompt information and/or second prompt information) in the status bar, and fig. 4 b-4 c take the first prompt information as an example, that is, the prompt information is used for prompting that the audio is turned off. The first prompt information may be an icon of a mute audio output function displayed in the status bar, the intelligent terminal may further display the first prompt information in a first preset floating window, for example, as shown in the floating window on the left side of fig. 4b, or as shown in the floating window on the right side of fig. 4b, the play interface of the target video may further display a prompt that the volume is changed to be mute, for example, as shown in the interface in fig. 4b, the prompt that the volume value is empty is displayed. Optionally, the intelligent terminal may further display the prompt message in a preset area, as shown in the left side of fig. 4c, and display the prompt message in a display area above the play window of the target video. Optionally, the first prompting message may be used to prompt the user that silent play is about to occur, such as "enter silent play after 5 seconds" shown in fig. 4b and 4 c.
Optionally, the first prompt message may be periodically displayed for a first preset time duration. That is, the prompt information is displayed in a flashing manner, may be displayed in a preset area in a flashing manner, and may also be displayed in a floating window in a flashing manner, which is not limited in the present application. Optionally, the prompt information may also be displayed in a preset area, and the display duration is a second preset duration, that is, the display duration disappears after the display. For example, after a period of time as shown in fig. 4 c. Alternatively, a notification that the playback is being muted may also be displayed in the preset notification bar. Optionally, the first prompting message may be used to prompt the user to play silently, that is, during the process of playing silently the audio included in the target video, the relevant prompt may be output. The prompt is the same as the output mode of the first prompt message, and details are not repeated here.
Similarly, for the prompt message for the intelligent terminal to enter the audio playing from the mute playing target audio, as shown in fig. 4d to 4e, fig. 4d and 4e are schematic user interface diagrams of the information processing method according to the first embodiment. Fig. 4d and 4e also illustrate the target audio corresponding to the target image frame, that is, the target audio is the audio included in the target video. The output mode of the prompt message may be the same as the output mode of fig. 4b and 4 c. Optionally, the volume value of the resume play displayed on the right side in the video play interface in fig. 4d and 4 e. Alternatively, as shown on the right side of fig. 4e, deleting the notification prompting silent play in the pull-down notification bar may indicate that voiced play is about to be entered, that the volume of the play is not zero, or that voiced play is occurring.
In the method described in fig. 3, the intelligent terminal obtains the reference environment audio by obtaining data of the target audio; and the target audio and the reference environment audio meet preset conditions, the target audio is output in a mute mode, and preset information is displayed. Therefore, based on the method described in fig. 3, the currently played target audio can be silently played in a noisy environment, and the preset information of the played target audio is displayed, so that the user can still obtain the content played by the intelligent terminal in the noisy environment. The operation that the user needs to adjust the volume is also simplified, the adjusting effect is improved, and the user experience is improved.
Referring to fig. 5, fig. 5 is another flow chart illustrating an information display according to a second embodiment. The information processing method shown in FIG. 5 includes steps 501 to 502. The method of the embodiment of the present application may be executed by the intelligent terminal shown in fig. 1, or may be executed by a chip in the intelligent terminal, and the intelligent terminal may be applied to the communication network system shown in fig. 2. The method execution subject shown in fig. 5 takes an intelligent terminal as an example.
It should be noted that the same or similar parts between the various embodiments in this application may be referred to each other. In the embodiments and the implementation methods/implementation methods in the embodiments in the present application, unless otherwise specified or conflicting in logic, terms and/or descriptions between different embodiments and between various implementation methods/implementation methods in various embodiments have consistency and can be mutually cited, and technical features in different embodiments and various implementation methods/implementation methods in various embodiments can be combined to form new embodiments, implementation methods, or implementation methods according to the inherent logic relationships thereof. The above-described embodiments of the present application do not limit the scope of the present application. Wherein:
501. and acquiring a preset style of text information and/or an image corresponding to the target audio.
Optionally, when the target audio and the reference environmental audio satisfy a mute output condition, the intelligent terminal may play the target audio in a mute manner, and the display information output according to the preset rule may include text information and/or images corresponding to the target audio, so as to provide subtitles for the user in a subtitle-free scene. Optionally, the intelligent terminal may determine or generate the target audio as an audio played by the target application, identify whether text information and/or an image corresponding to the target audio exists on an interface of the target audio, obtain an identification result, and output the display information according to the type and/or the identification result of the target application.
Optionally, the target audio is played by running the target application on the intelligent terminal, so that the output content can be determined according to the type and the identification result of the target application. Optionally, when the target application is determined to be of the first type, the target application may be chat software, the target application may provide a function of voice chat communication, and the target application is applied to a scenario of voice chat and/or video chat, the intelligent terminal may output the first prompt message and/or the second prompt message when the target audio is played in a mute manner. Optionally, when it is determined that the target application is of the second type and the identified type is no, the target application may be a video-type application, the target application may also be an application of a browser, and the target application may also be an application capable of providing a video playing function, which is not limited in this application. In the case where the video is played in the target application, and it is recognized that there is no text information and/or image corresponding to the target audio, that is, the video is played without any subtitle or image, and thus, the text information and/or image corresponding to the first prompt information and the target audio may be output. Alternatively, in a case where it is determined that the target application is the audio of the second type application and the recognition result is yes, only the first prompt information may be output. That is, when it is detected that the target audio is the audio of the video output by the target application and the corresponding text information and/or image exists in the video, the text information and/or image corresponding to the target audio does not need to be output again, and only the prompt information needs to be output.
Optionally, the method includes acquiring text information and/or an image corresponding to the target audio at the intelligent terminal, and acquiring a preset style corresponding to the text information and the image, where the preset style may be a display style of the text information, and includes at least one of a display position, a display mode, and a display style of the text information. And determining text information having a preset pattern and/or an image having a preset pattern as the display information. Optionally, the intelligent terminal may obtain the text information corresponding to the target audio conversion by obtaining a subtitle file of the target audio, for example, obtaining a subtitle file of a target video including the target audio. Optionally, the intelligent terminal may input the target audio into a neural network of a threshold, and obtain text information of the target audio through speech recognition, which is not limited in this application. Optionally, the image corresponding to the target audio may be an image obtained in association with the target audio and/or the target video, or an image obtained according to the recognized text information, which is not limited in this application.
502. And outputting the text information and/or the image corresponding to the target audio in the preset mode under the condition of outputting the target image frame.
Alternatively, in the case where the target audio is not the audio included in the target video, text information and/or images having a preset pattern may be determined as the display information. Alternatively, in a case where the target audio is audio included in the target video, in a case where the target image frame of the target video is output, the text information of the preset style is output. Optionally, the intelligent terminal may further receive an adjustment operation for adjusting the text information and/or the image after displaying the display information under the condition of playing the target audio in a mute manner, and adjust the display mode of the displayed text information and/or image according to the adjustment operation to obtain updated text information and/or image.
Optionally, the adjustment operation may include, but is not limited to, at least one of: the method comprises the steps of adjusting the display position of the text information and/or the image, adjusting the display style of the text information and/or the image, triggering an adjusting control carried by the text information and/or the image, closing the display of the text information in the text information and/or the image, and adjusting the display mode of the text information and/or the image to be displayed in a second preset floating window. Other adjustment operations may also be included, which are not limited in this application.
Referring to fig. 6a to 6d, fig. 6a to 6d are schematic views of a user interface for an adjustment operation according to a second embodiment. Fig. 6a to 6d explain the target audio corresponding to the target image frame, that is, the target audio is the audio included in the target video, and the display information includes text information. As shown in fig. 6a, in the case of outputting the target image frame, the user may drag the text information to adjust the display position of the text information, and the display position of the text information may be within the display area of the target video or outside the display area of the target video.
Optionally, as shown in fig. 6b, the user may click a setting control in the adjustment control to adjust the display style of the text information, and as shown in the right side of fig. 6b, after receiving a trigger operation of the user for the setting control, the intelligent terminal may display a prompt box, where the prompt box includes a color and a font size for adjusting the text information, so that the text information is not obscured from the user when the color of the text information is similar to that of the image frame of the target video. As shown in fig. 6c, the user may click the floating window control, the intelligent terminal receives an adjustment operation of the user for the floating window control, and adjusts the display mode of the text information to be displayed in the second preset floating window, as shown on the right side of fig. 6 c.
Optionally, fig. 6c is only an example, and text information and/or a display mode may be adjusted to be displayed in a floating window in other ways, which is not limited in this application. As shown in fig. 6d, the user may click a closing control of the adjustment control of the text information, and the intelligent terminal closes and displays the text information in response to a triggering operation of the closing control. For example, when the target video includes the embedded subtitles, that is, the subtitles are included in the image frames of the target video, the text information does not need to be generated, and the user can turn off the display processing of the generated text information. Optionally, the descriptions shown in fig. 6a to fig. 6d only take text information as an example for explanation, and the same is also true for images in the display information, which is not repeated here, and the design may be performed according to the user interface requirement of the target application, which is not limited in this application.
Referring to fig. 6e to fig. 6g, fig. 6e and fig. 6g are schematic views of a user interface for an adjustment operation according to a second embodiment. Fig. 6 e-6 g illustrate the target audio being audio only, and the display information includes text information and/or images. As shown in fig. 6e, in the target audio playing interface, the intelligent terminal may receive an adjustment operation of the user for a display position of the text information, for example, a dragging operation shown in fig. 6e, and the target audio playing interface may include a display area for displaying the text information, and the user may drag the text information to display the display area.
Optionally, as shown in fig. 6f, the intelligent terminal may receive a trigger operation for an adjustment control carried by the text information and/or the image, where the trigger operation may be for a floating window, and adjust a display mode of the text information and/or the image to be displayed in a second preset floating window. Optionally, fig. 6f is also only an example, and the display manners of the text information and/or the image may also be adjusted to be displayed in the floating window in other manners, which is not limited in this application.
Optionally, as shown in fig. 6g, the intelligent terminal may receive a trigger operation for closing a control in an adjustment control carried by the text information and/or the image, and perform display closing processing on the text information and/or the image. For example, when the target audio is music, the text information may be lyrics of the music, and the image may be an image related to the music. When the lyrics of the target audio have been displayed in the user interface playing the target audio, the user may turn off displaying the text information and/or image. Optionally, the descriptions shown in fig. 6e to fig. 6g are only for the example of text information, and the same is also true for the images in the display information, which is not repeated here, and the design may be performed according to the user interface requirement of the target application, which is not limited in this application.
In the method described in fig. 5, the intelligent terminal outputs the text information and/or the image corresponding to the target audio in the preset pattern by acquiring the preset pattern of the text information and/or the image corresponding to the target audio in the case of outputting the target image frame. Therefore, based on the method described in fig. 5, the user can acquire the content of the target audio or the target video under the condition of playing the audio in a mute manner, so that the adjustment operation of the user for the volume is simplified, the user can acquire the content of the played audio in a noisy environment, the adjustment effect is improved, and the user experience is further improved.
Referring to fig. 7, fig. 7 is a flowchart illustrating an information processing method according to a third embodiment. The information processing method shown in FIG. 7 includes 701 to 702. The method of the embodiment of the present application may be executed by the intelligent terminal shown in fig. 1, or may be executed by a chip in the intelligent terminal, and the intelligent terminal may be applied to the communication network system shown in fig. 2. The method shown in fig. 7 is executed by taking an intelligent terminal as an example.
It should be noted that the same or similar parts between the various embodiments in this application may be referred to each other. In the embodiments and the implementation methods/implementation methods in the embodiments in the present application, unless otherwise specified or conflicting in logic, terms and/or descriptions between different embodiments and between various implementation methods/implementation methods in various embodiments have consistency and can be mutually cited, and technical features in different embodiments and various implementation methods/implementation methods in various embodiments can be combined to form new embodiments, implementation methods, or implementation methods according to the inherent logic relationships thereof. The above-described embodiments of the present application do not limit the scope of the present application. Wherein:
701. and receiving the opening operation of the mute audio output function.
702. And responding to the starting operation, recording the played target audio, wherein the playing volume of the target audio is not zero, and obtaining the reference environment audio.
In a possible implementation manner, the intelligent terminal may receive a start operation of the mute audio output function, and in response to the start operation, the intelligent terminal acquires data of the target audio when the target audio is played through the speaker, that is, records and judges the data. Alternatively, the turning on operation of the mute output audio function may include, but is not limited to, one of: the method comprises the steps of triggering operation aiming at a preset setting menu, triggering operation aiming at a preset control included in a notification bar, and triggering operation aiming at a setting menu in a target application, wherein the target audio is the audio in the target application and the triggering operation aiming at a preset function menu, and the function menu is used for switching the preset function.
Referring to fig. 8 a-8 d together, fig. 8 a-8 d are schematic diagrams illustrating a user interface for turning on a mute output audio function according to a third embodiment. As shown in fig. 8a, the notification bar may be a menu of mute output in the notification bar displayed after the status bar is pulled down, the user may click the icon for mute play to trigger operation, and the smart terminal receives the triggering operation of the icon for mute play in the notification bar and starts the function of mute play audio. As shown in fig. 8b and 8c, the preset setting menu may be a mute-to-play menu in a shortcut menu, and the shortcut menu may be initiated by a slide operation, as shown in fig. 8 b. Alternatively, it may be initiated by triggering a floating window. For example, the user may click on the hover ball and display a quick menu. As shown in the left side of fig. 8c, the preset setting menu may also be in the system setting menu, i.e. the function of playing audio in mute may be a system level function. As shown on the right side of fig. 8c, the opening operation may be a trigger operation for a preset function menu, which may be used to switch a preset function. Fig. 8c illustrates an energy saving function, which may be a selection for power consumption or performance, but is not limited in this application. It can be understood that after the target audio is played in a mute manner, the power consumption can be saved, so that the purposes of saving energy and prolonging the standby time of the battery are achieved. As shown in fig. 8d, the function of playing audio in a mute manner is a function of an application program, the application program may play a video including a target audio, or may play the target audio, and the opening operation may be a trigger operation of a setting menu in the target application. The left side of fig. 8d is a user interface for playing a target video by an application program, and the function of playing audio in a mute manner is turned on for the video, and the right side of fig. 8d is a user interface for playing a target audio by an application program, and the function of playing audio in a mute manner is turned on for the audio.
In the method depicted in fig. 7, the smart terminal performs an on operation by receiving a mute output audio function; and responding to the starting operation, recording the played target audio, wherein the playing volume of the target audio is not zero, and obtaining the reference environment audio. Therefore, based on the method described in fig. 7, the user can conveniently and flexibly turn on the function of playing the audio in a mute manner, the operation of turning on the audio in the mute manner by the user is simplified, the user is ensured to obtain the content of the target audio played through the loudspeaker in a noisy environment, the user experience is improved, and the applicability of the intelligent terminal is also improved.
Referring to fig. 9, fig. 9 is a schematic structural diagram of an information processing apparatus 90 according to an embodiment of the present application, where the information processing apparatus 90 includes an obtaining unit 901, a playing unit 902, a separating unit 903, a processing unit 904, an output unit 905, a receiving unit 906, and an adjusting unit 907, where:
an obtaining unit 901, configured to obtain data of a target audio to obtain a reference environment audio;
the playing unit 902 is configured to play the target audio and the reference environmental audio in a mute manner, and output preset information, when the target audio and the reference environmental audio satisfy a preset condition.
Optionally, the playing unit 902 is configured to mute and play the target audio and the reference environmental audio according to a preset condition, and output preset information, and specifically configured to:
and the target audio and the reference environment audio meet the condition of mute audio output, the target audio is played in a mute manner, and display information is output according to a preset rule.
Optionally, the display information includes at least one of:
first prompt information for prompting audio closing;
second prompting information used for prompting the audio-to-text information;
and text information and/or images corresponding to the target audio.
Optionally, the playing unit 902 is configured to output display information according to a preset rule, and specifically configured to:
determining or generating the target audio as the audio of the target application;
identifying whether text information and/or images corresponding to the target audio exist in an interface for playing the target audio to obtain an identification result;
and outputting the display information according to the type of the target application and/or the identification result.
Optionally, the playing unit 902 is configured to output the display information according to the type of the target application and the identification result, where the display information includes at least one of the following:
outputting the first prompt information and/or the second prompt information when the target application is of a first type;
outputting text information and/or images corresponding to the first prompt information and the target audio when the target audio is of a second type and the identification result is negative;
and outputting the first prompt information when the target audio is the audio of the second type application and the identification result is yes.
Optionally, the information processing apparatus 90 further includes:
a separating unit 903, configured to separate first environment information from the reference environment audio according to the target audio to obtain a first environment audio;
a processing unit 904, configured to determine or generate first analog loudness information according to the target audio, and determine or generate second analog loudness information according to the first environmental audio;
the processing unit 904 is further configured to determine or generate a condition that satisfies the mute output audio when the first analog loudness information and/or the second analog loudness information satisfies a preset environmental condition.
Optionally, the processing unit 904 is configured to determine or generate first analog loudness information according to the target audio, and determine or generate second analog loudness information according to the first environmental audio, and specifically configured to:
acquiring a first audio parameter of the target audio and a second audio parameter of the first environment audio;
determining or generating the first simulated loudness information of the first audio parameter according to a first corresponding relationship between the simulated loudness value and the audio parameter;
and determining or generating the second simulated loudness information of the second audio parameter according to the second corresponding relationship between the simulated loudness value and the audio parameter.
Optionally, the obtaining unit 901 is further configured to obtain second environment information to obtain a second environment audio;
the processing unit 904 is further configured to determine or generate third analog loudness information according to the second ambient audio;
the processing unit 904 is further configured to determine or generate that the audio playing condition is satisfied when the first analog loudness information and/or the third analog loudness value satisfy a preset audio condition;
an output unit 905 is configured to output the sound of the target audio.
Optionally, the output mode of the first prompt message and/or the second prompt message includes at least one of the following:
displaying the first prompt message and/or the second prompt message in a status bar;
displaying the first prompt message and/or the second prompt message in a first preset floating window;
displaying the first prompt message and/or the second prompt message in a preset area;
periodically displaying the first prompt message and/or the second prompt message with a first preset time length;
and displaying the first prompt message and/or the second prompt message in the preset area, wherein the duration of the first prompt message and/or the second prompt message is a second preset duration.
Optionally, the target audio corresponds to a target image frame, and the playing unit 902 is configured to output text information and/or an image corresponding to the target audio, and specifically configured to:
acquiring text information corresponding to the target audio and a preset pattern of the image;
and outputting the text information and/or the image corresponding to the target audio in the preset mode under the condition of outputting the target image frame.
Optionally, the obtaining unit 901 is configured to obtain data of a target audio to obtain a reference environmental audio, and specifically configured to:
receiving the starting operation of the mute audio output function;
responding to the starting operation, recording the played target audio, wherein the playing volume of the target audio is not zero, and obtaining the reference environment audio.
Optionally, the information processing apparatus 90 further includes:
a receiving unit 906, configured to receive an adjustment operation for adjusting text information and/or an image corresponding to the target audio;
an adjusting unit 907, configured to adjust a display manner of the text information and/or the image corresponding to the target audio to obtain updated text information and/or an updated image.
Optionally, the adjusting operation includes at least one of:
an adjustment operation for a display position of the text information and/or the image;
adjusting the display style of the text information and/or the image;
triggering operation aiming at the text information and/or the adjusting control carried by the image;
closing display operation for the text information and/or the text information in the image;
and adjusting the display mode of the text information and/or the image to be displayed in a second preset floating window.
The embodiment of the present application further provides an intelligent terminal, where the intelligent terminal includes a memory and a processor, where the memory stores an information processing program, and the information processing program, when executed by the processor, implements the steps of the information processing method in any of the above embodiments.
The embodiment of the present application further provides a computer-readable storage medium, where an information processing program is stored on the storage medium, and when the information processing program is executed by a processor, the steps of the information processing method in any of the above embodiments are implemented.
In the embodiments of the intelligent terminal and the computer-readable storage medium provided in the present application, all technical features of any one of the embodiments of the information processing method may be included, and the expanding and explaining contents of the specification are basically the same as those of the embodiments of the method, and are not described herein again.
Embodiments of the present application also provide a computer program product, which includes computer program code, when the computer program code runs on a computer, the computer is caused to execute the method in the above various possible embodiments.
Embodiments of the present application further provide a chip, which includes a memory and a processor, where the memory is used to store a computer program, and the processor is used to call and run the computer program from the memory, so that a device in which the chip is installed executes the method in the above various possible embodiments.
It is to be understood that the foregoing scenarios are only examples, and do not constitute a limitation on application scenarios of the technical solutions provided in the embodiments of the present application, and the technical solutions of the present application may also be applied to other scenarios. For example, as can be known by those skilled in the art, with the evolution of system architecture and the emergence of new service scenarios, the technical solution provided in the embodiments of the present application is also applicable to similar technical problems.
The above-mentioned serial numbers of the embodiments of the present application are merely for description, and do not represent the advantages and disadvantages of the embodiments.
The steps in the method of the embodiment of the application can be sequentially adjusted, combined and deleted according to actual needs.
The units in the device of the embodiment of the application can be combined, divided and deleted according to actual needs.
In the present application, the same or similar term concepts, technical solutions and/or application scenario descriptions will be generally described only in detail at the first occurrence, and when the description is repeated later, the detailed description will not be repeated in general for brevity, and when understanding the technical solutions and the like of the present application, reference may be made to the related detailed description before the description for the same or similar term concepts, technical solutions and/or application scenario descriptions and the like which are not described in detail later.
In the present application, each embodiment is described with emphasis, and reference may be made to the description of other embodiments for parts that are not described or illustrated in any embodiment.
All possible combinations of the technical features in the embodiments are not described in the present application for the sake of brevity, but should be considered as the scope of the present application as long as there is no contradiction between the combinations of the technical features.
Through the above description of the embodiments, those skilled in the art will clearly understand that the method of the above embodiments can be implemented by software plus a necessary general hardware platform, and certainly can also be implemented by hardware, but in many cases, the former is a better implementation manner. Based on such understanding, the technical solutions of the present application may be embodied in the form of a software product, which is stored in a storage medium (e.g., ROM/RAM, magnetic disk, optical disk) and includes instructions for enabling a terminal device (e.g., a mobile phone, a computer, a server, a controlled terminal, or a network device) to execute the method of each embodiment of the present application.
In the above embodiments, the implementation may be wholly or partially realized by software, hardware, firmware, or any combination thereof. When implemented in software, may be implemented in whole or in part in the form of a computer program product. The computer program product includes one or more computer instructions. The procedures or functions according to the embodiments of the present application are all or partially generated when the computer program instructions are loaded and executed on a computer. The computer may be a general purpose computer, a special purpose computer, a network of computers, or other programmable device. The computer instructions may be stored on a computer readable storage medium or transmitted from one computer readable storage medium to another, for example, the computer instructions may be transmitted from one website, computer, server, or data center to another website, computer, server, or data center by wire (e.g., coaxial cable, fiber optic, digital subscriber line) or wirelessly (e.g., infrared, wireless, microwave, etc.). The computer-readable storage medium can be any available medium that can be accessed by a computer or a data storage device, such as a server, a data center, etc., that incorporates one or more of the available media. The usable medium may be a magnetic medium (e.g., floppy Disk, memory Disk, magnetic tape), an optical medium (e.g., DVD), or a semiconductor medium (e.g., Solid State Disk (SSD)), among others.
The above description is only a preferred embodiment of the present application, and not intended to limit the scope of the present application, and all modifications of equivalent structures and equivalent processes, which are made by the contents of the specification and the drawings of the present application, or which are directly or indirectly applied to other related technical fields, are included in the scope of the present application.

Claims (10)

1. An information processing method characterized by comprising the steps of:
s10, acquiring data of the target audio to obtain a reference environment audio;
and S20, when the target audio and the reference environment audio meet preset conditions, playing the target audio in a mute mode, and outputting preset information.
2. The method of claim 1, wherein the step of S20 includes:
and the target audio and the reference environment audio meet the condition of mute audio output, the target audio is played in a mute mode, and display information is output according to a preset rule.
3. The method according to claim 2, wherein the outputting the display information according to the preset rule comprises:
determining or generating the target audio as the audio of the target application;
identifying whether text information and/or images corresponding to the target audio exist in an interface for playing the target audio to obtain an identification result;
and outputting the display information according to the type of the target application and/or the identification result.
4. The method of claim 3, wherein the outputting the display information according to the type of the target application and the recognition result comprises at least one of:
outputting the first prompt message and/or the second prompt message under the condition that the target application is of a first type;
under the condition that the target audio is of a second type and the recognition result is negative, outputting text information and/or images corresponding to the first prompt information and the target audio;
and if the target audio is the audio of the second type of application and the identification result is yes, outputting the first prompt information.
5. The method according to any one of claims 2 to 4, wherein before the step of S20, the method further comprises:
separating first environment information from the reference environment audio according to the target audio to obtain a first environment audio;
determining or generating first simulated loudness information according to the target audio, and determining or generating second simulated loudness information according to the first environmental audio;
and when the first analog loudness information and/or the second analog loudness information meet preset environmental conditions, determining or generating conditions meeting the mute output audio.
6. The method of claim 5, wherein determining or generating first simulated loudness information from the target audio and second simulated loudness information from the first ambient audio comprises:
acquiring a first audio parameter of the target audio and a second audio parameter of the first environment audio;
determining or generating the first simulated loudness information of the first audio parameter according to a first corresponding relationship between simulated loudness information and audio parameters;
and determining or generating the second simulated loudness information of the second audio parameter according to the second corresponding relation between the simulated loudness information and the audio parameter.
7. The method according to claim 6, wherein after the step of S20, the method further comprises:
acquiring second environment information to obtain second environment audio;
determining or generating third simulated loudness information according to the second ambient audio;
when the first analog loudness information and/or the third analog loudness value meet a preset audio condition, determining or generating that the audio playing condition is met;
outputting the sound of the target audio.
8. The method according to any one of claims 2 to 4, wherein when the target audio and the reference environmental audio satisfy a preset condition, the target audio is played in a mute manner, and after preset information is output, the method further comprises:
receiving an adjusting operation for adjusting text information and/or images corresponding to the target audio;
and adjusting the display mode of the text information and/or the image corresponding to the target audio to obtain the updated text information and/or image.
9. An intelligent terminal, comprising: memory, processor, wherein the memory has stored thereon an information processing program which, when executed by the processor, implements the steps of the information processing method according to any one of claims 1 to 8.
10. A computer-readable storage medium, characterized in that the storage medium has stored thereon a computer program which, when being executed by a processor, realizes the steps of the information processing method according to any one of claims 1 to 8.
CN202210655732.9A 2022-06-10 2022-06-10 Information processing method, intelligent terminal and storage medium Pending CN115118812A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210655732.9A CN115118812A (en) 2022-06-10 2022-06-10 Information processing method, intelligent terminal and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210655732.9A CN115118812A (en) 2022-06-10 2022-06-10 Information processing method, intelligent terminal and storage medium

Publications (1)

Publication Number Publication Date
CN115118812A true CN115118812A (en) 2022-09-27

Family

ID=83326927

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210655732.9A Pending CN115118812A (en) 2022-06-10 2022-06-10 Information processing method, intelligent terminal and storage medium

Country Status (1)

Country Link
CN (1) CN115118812A (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2008109595A (en) * 2006-10-27 2008-05-08 Nec Corp Automatic speech method switching system, server, method, and program
JP2012095047A (en) * 2010-10-26 2012-05-17 Panasonic Corp Speech processing unit
CN105898573A (en) * 2016-05-03 2016-08-24 北京小米移动软件有限公司 Method and device for multimedia file playing
CN106331371A (en) * 2016-09-14 2017-01-11 维沃移动通信有限公司 Volume adjustment method and mobile terminal
CN108307043A (en) * 2017-12-28 2018-07-20 努比亚技术有限公司 Speech message conversion method, mobile terminal and computer readable storage medium
CN111107283A (en) * 2019-11-28 2020-05-05 联想(北京)有限公司 Information display method, electronic equipment and storage medium

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2008109595A (en) * 2006-10-27 2008-05-08 Nec Corp Automatic speech method switching system, server, method, and program
JP2012095047A (en) * 2010-10-26 2012-05-17 Panasonic Corp Speech processing unit
CN105898573A (en) * 2016-05-03 2016-08-24 北京小米移动软件有限公司 Method and device for multimedia file playing
CN106331371A (en) * 2016-09-14 2017-01-11 维沃移动通信有限公司 Volume adjustment method and mobile terminal
CN108307043A (en) * 2017-12-28 2018-07-20 努比亚技术有限公司 Speech message conversion method, mobile terminal and computer readable storage medium
CN111107283A (en) * 2019-11-28 2020-05-05 联想(北京)有限公司 Information display method, electronic equipment and storage medium

Similar Documents

Publication Publication Date Title
CN107231473B (en) Audio output regulation and control method, equipment and computer readable storage medium
CN107463243B (en) Screen control method, mobile terminal and computer readable storage medium
CN108196750A (en) The method, apparatus and computer readable storage medium of screen interface sectional drawing
CN110187808B (en) Dynamic wallpaper setting method and device and computer-readable storage medium
CN112004174A (en) Noise reduction control method and device and computer readable storage medium
CN108322602B (en) Method, terminal and computer readable storage medium for processing application no response
CN108052304B (en) Volume adjusting method, volume adjusting equipment and computer storage medium
CN109729267A (en) Filter selection method, mobile terminal and computer readable storage medium
CN107728990B (en) Audio playing method, mobile terminal and computer readable storage medium
CN113176869A (en) Screen-projecting audio time delay control method and device and computer readable storage medium
CN112712815A (en) Software-based pop sound suppression method, terminal and computer readable medium
CN109561221B (en) Call control method, device and computer readable storage medium
CN108959382B (en) Audio and video detection method and mobile terminal
CN107678622B (en) Application icon display method, terminal and storage medium
CN115955529A (en) Screen recording method, intelligent terminal and storage medium
CN115914719A (en) Screen projection display method, intelligent terminal and storage medium
CN112532787B (en) Earphone audio data processing method, mobile terminal and computer readable storage medium
CN109005281A (en) A kind of In Call adjusting method, mobile terminal and computer readable storage medium
CN115469949A (en) Information display method, intelligent terminal and storage medium
CN114967916A (en) Control method, intelligent terminal and storage medium
CN114666440A (en) Application program control method, intelligent terminal and storage medium
CN113805837A (en) Audio processing method, mobile terminal and storage medium
CN114065168A (en) Information processing method, intelligent terminal and storage medium
CN110278402B (en) Dual-channel audio processing method and device and computer readable storage medium
CN113329347A (en) Processing method of Bluetooth device, mobile terminal and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination