CN115955529A - Screen recording method, intelligent terminal and storage medium - Google Patents
Screen recording method, intelligent terminal and storage medium Download PDFInfo
- Publication number
- CN115955529A CN115955529A CN202211570405.XA CN202211570405A CN115955529A CN 115955529 A CN115955529 A CN 115955529A CN 202211570405 A CN202211570405 A CN 202211570405A CN 115955529 A CN115955529 A CN 115955529A
- Authority
- CN
- China
- Prior art keywords
- screen recording
- audio data
- recording
- audio
- screen
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Images
Landscapes
- Telephone Function (AREA)
Abstract
The application provides a screen recording method, an intelligent terminal and a storage medium, wherein the screen recording method is applied to the intelligent terminal and comprises the following steps: responding to a screen recording instruction, and displaying a selection control on a screen recording interface; receiving a trigger operation aiming at the selection control, and determining or generating an audio processing mode based on the trigger operation; and when the audio processing mode is noise reduction processing, performing noise reduction processing on the audio data corresponding to the screen recording process according to the target voiceprint. According to the method and the device, when the audio processing mode is determined to be noise reduction processing, noise reduction processing is performed on the audio data according to the target voiceprint, so that other audio data except the audio data corresponding to the target voiceprint are not recorded in the screen recording process, the recorded file does not contain other audio data except the audio data corresponding to the target voiceprint, the problem that the recorded file is mixed with environmental noise, the user cannot know the required sound information from the mixed audio and video is solved, and user experience is improved.
Description
Technical Field
The application relates to the technical field of terminals, in particular to a screen recording method, an intelligent terminal and a storage medium.
Background
With the continuous development of terminal technology, in order to better meet the terminal use experience of a user, the terminal provides a screen recording function for the user, and the user can record videos in the terminal operation process by using the screen recording function.
In the course of conceiving and implementing the present application, the inventors found that at least the following problems existed: in the process of recording the screen, the existing screen recording can directly start recording, and environmental sound or noise can be recorded in the screen recording process, so that the recording file is mixed with the environmental noise, and a user cannot know the wanted sound information from the mixed audio and video.
The foregoing description is provided for general background information and is not admitted to be prior art.
Disclosure of Invention
In order to solve the technical problems, the present application provides a screen recording method, an intelligent terminal and a storage medium, so that when a user records a screen, a recording file that does not contain other audio data except for the required audio data is generated, and the operation is simple and convenient.
In order to solve the technical problem, the present application provides a screen recording method, which can be applied to an intelligent terminal, and includes:
s10: responding to a screen recording instruction, and displaying a selection control on a screen recording interface;
s20: receiving a trigger operation aiming at the selection control, and determining or generating an audio processing mode based on the trigger operation;
s30: and when the audio processing mode is noise reduction processing, performing noise reduction processing on the audio data corresponding to the screen recording process according to the target voiceprint.
Optionally, the step S30 includes:
when the audio processing mode is noise reduction processing, displaying a voice input window in a screen recording interface;
receiving a voice signal input aiming at the voice input window, and acquiring the target voiceprint based on the voice signal;
and performing noise reduction processing on the audio data corresponding to the screen recording process according to the target voiceprint.
Optionally, after the step of determining the audio processing manner based on the trigger operation, the method further includes:
s21, after a screen recording process is started based on the audio processing mode, displaying an audio recording control in the screen recording interface;
and S22, receiving a first operation aiming at the audio recording control, and interrupting recording of audio data.
Optionally, after the step S22, the method further includes:
and responding to the starting operation aiming at the audio recording control, and continuing to record the audio data.
Optionally, after the step S20, the method further includes:
when the audio processing mode is non-noise reduction processing, acquiring volume values of other audio data except the target audio data corresponding to the target voiceprint;
and when the volume value is greater than or equal to the preset volume value, displaying a selection control on a recording screen interface and/or displaying an audio recording control.
Optionally, the step S10 includes:
responding to the screen recording instruction, and outputting a screen recording mode selection list;
and receiving a selection operation aiming at the screen recording mode selection list, and displaying the selection control on a screen recording interface.
Optionally, after the step S30, the method further includes:
receiving a screen recording stopping instruction aiming at the screen recording process, and determining or generating a recording file according to the processed audio data and the video data corresponding to the screen recording process;
determining or generating an audio timeline from the audio data;
and playing the recording file in a preview interface, and displaying the audio time axis in an overlapping manner on the preview interface.
Optionally, the step of playing the recording file in the preview interface further includes:
acquiring a current playing time point in the video data playing process, or acquiring a selected time point aiming at the audio time axis; taking the current playing time point or the selected time point as a preview time point;
acquiring audio data matched with the preview time point, and determining or generating target information according to the audio data; and overlaying the target information in a target video frame corresponding to the preview time point, and displaying the overlaid target image frame.
The application also provides an intelligent terminal, including: the device comprises a memory and a processor, wherein the memory stores a screen recording program, and the screen recording program realizes the steps of the method when being executed by the processor.
The present application also provides a storage medium storing a computer program which, when executed by a processor, performs the steps of the method as described above.
As described above, the screen recording method of the present application may be applied to an intelligent terminal, and includes the steps of: s10: responding to a screen recording instruction, and displaying a selection control on a screen recording interface; s20: receiving a trigger operation aiming at the selection control, and determining or generating an audio processing mode based on the trigger operation; s30: and when the audio processing mode is noise reduction processing, performing noise reduction processing on the audio data corresponding to the screen recording process according to the target voiceprint. Through the technical scheme, the method and the device can realize that other audio data except the audio data corresponding to the target voiceprint are not recorded when the screen recording process is executed, thereby generating the function of not containing other audio data except the audio data corresponding to the target voiceprint, solving the problem that the recorded file is mixed with environmental noise because the environmental sound or noise can be recorded simultaneously in the screen recording process, and further improving the user experience.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the application and, together with the description, serve to explain the principles of the application. In order to more clearly illustrate the technical solutions of the embodiments of the present application, the drawings required to be used in the description of the embodiments will be briefly described below, and it is obvious for those skilled in the art to obtain other drawings without inventive step.
Fig. 1 is a schematic diagram of a hardware structure of an intelligent terminal implementing various embodiments of the present application;
fig. 2 is a diagram illustrating a communication network system architecture according to an embodiment of the present application;
fig. 3 is a flowchart illustrating a screen recording method according to the first embodiment;
fig. 4 is a detailed flowchart illustrating the step S10 of the screen recording method according to the first embodiment;
fig. 5 is a detailed flowchart of step S30 of the screen recording method according to the first embodiment;
fig. 6 is a flowchart illustrating a screen recording method according to the second embodiment;
fig. 7 is a flowchart illustrating a screen recording method according to the second embodiment;
fig. 8 is a flowchart illustrating a screen recording method according to the third embodiment;
fig. 9 is a flowchart illustrating a screen recording method according to the fourth embodiment;
fig. 10 is a screen recording interface diagram of a screen recording method according to the fourth embodiment;
fig. 11 is an audio timeline diagram of a screen recording method according to the fourth embodiment;
fig. 12 is a detailed flowchart of step S70 of the screen recording method according to the fourth embodiment.
The implementation, functional features and advantages of the object of the present application will be further explained with reference to the embodiments, and with reference to the accompanying drawings. With the above figures, there are shown specific embodiments of the present application, which will be described in more detail below. These drawings and written description are not intended to limit the scope of the inventive concepts in any manner, but rather to illustrate the inventive concepts to those skilled in the art by reference to specific embodiments.
Detailed Description
Reference will now be made in detail to the exemplary embodiments, examples of which are illustrated in the accompanying drawings. When the following description refers to the accompanying drawings, like numbers in different drawings represent the same or similar elements unless otherwise indicated. The implementations described in the following exemplary examples do not represent all implementations consistent with the present application. Rather, they are merely examples of apparatus and methods consistent with certain aspects of the present application, as detailed in the appended claims.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising a … …" does not exclude the presence of additional like elements in a process, method, article, or apparatus that comprises the element, and further, components, features, elements, and/or steps that may be similarly named in various embodiments of the application may or may not have the same meaning, unless otherwise specified by its interpretation in the embodiment or by context with further embodiments.
It should be understood that although the terms first, second, third, etc. may be used herein to describe various information, such information should not be limited to these terms. These terms are only used to distinguish one type of information from another. For example, first information may also be referred to as second information, and similarly, second information may also be referred to as first information, without departing from the scope herein. The word "if," as used herein, may be interpreted as "at … …" or "at … …" or "in response to a determination," depending on the context. Also, as used herein, the singular forms "a", "an" and "the" are intended to include the plural forms as well, unless the context indicates otherwise. It will be further understood that the terms "comprises," "comprising," "includes" and/or "including," when used in this specification, specify the presence of stated features, steps, operations, elements, components, items, species, and/or groups, but do not preclude the presence, or addition of one or more other features, steps, operations, elements, components, items, species, and/or groups thereof. The terms "or," "and/or," "including at least one of the following," and the like, as used herein, are to be construed as inclusive or mean any one or any combination. For example, "includes at least one of: A. b, C "means" any of the following: a; b; c; a and B; a and C; b and C; a and B and C ", further for example," A, B or C "or" A, B and/or C "means" either: a; b; c; a and B; a and C; b and C; a and B and C'. An exception to this definition will occur only when a combination of elements, functions, steps or operations are inherently mutually exclusive in some way.
It should be understood that, although the steps in the flowcharts in the embodiments of the present application are shown in sequence as indicated by the arrows, the steps are not necessarily performed in sequence as indicated by the arrows. The steps are not performed in the exact order shown and may be performed in other orders unless explicitly stated herein. Moreover, at least some of the steps in the figures may include multiple sub-steps or multiple stages that are not necessarily performed at the same time, but may be performed at different times, in different orders, and may be performed alternately or at least partially with respect to other steps or sub-steps of other steps.
The words "if", as used herein, may be interpreted as "at … …" or "at … …" or "in response to a determination" or "in response to a detection", depending on the context. Similarly, the phrase "if determined" or "if detected (a stated condition or event)" may be interpreted as "upon determining" or "in response to determining" or "upon detecting (a stated condition or event)" or "in response to detecting (a stated condition or event)", depending on the context.
It should be noted that step numbers such as S10 and S20 are used herein for the purpose of more clearly and briefly describing the corresponding contents, and do not constitute a substantial limitation on the sequence, and those skilled in the art may perform S20 first and then S10 in the specific implementation, but these should be within the protection scope of the present application.
It should be understood that the specific embodiments described herein are merely illustrative of the present application and are not intended to limit the present application.
In the following description, suffixes such as "module", "component", or "unit" used to denote elements are used only for the convenience of description of the present application, and have no specific meaning in themselves. Thus, "module", "component" or "unit" may be used mixedly.
The smart terminal may be implemented in various forms. For example, the smart terminal described in the present application may include smart terminals such as a mobile phone, a tablet computer, a notebook computer, a palmtop computer, a Personal Digital Assistant (PDA), a Portable Media Player (PMP), a navigation device, a wearable device, a smart band, a pedometer, and the like, and fixed terminals such as a Digital TV, a desktop computer, and the like.
While the following description will be given by way of example of a smart terminal, those skilled in the art will appreciate that the configuration according to the embodiments of the present application can be applied to a fixed type terminal in addition to elements particularly used for mobile purposes.
Referring to fig. 1, which is a schematic diagram of a hardware structure of an intelligent terminal for implementing various embodiments of the present application, the intelligent terminal 100 may include: RF (Radio Frequency) unit 101, wiFi module 102, audio output unit 103, a/V (audio/video) input unit 104, sensor 105, display unit 106, user input unit 107, interface unit 108, memory 109, processor 110, and power supply 111. Those skilled in the art will appreciate that the intelligent terminal architecture shown in fig. 1 does not constitute a limitation of the intelligent terminal, and that the intelligent terminal may include more or fewer components than shown, or some components may be combined, or a different arrangement of components.
The following specifically describes each component of the intelligent terminal with reference to fig. 1:
the radio frequency unit 101 may be configured to receive and transmit signals during information transmission and reception or during a call, and specifically, receive downlink information of a base station and then process the downlink information to the processor 110; in addition, the uplink data is transmitted to the base station. Typically, radio frequency unit 101 includes, but is not limited to, an antenna, at least one amplifier, a transceiver, a coupler, a low noise amplifier, a duplexer, and the like. In addition, the radio frequency unit 101 can also communicate with a network and other devices through wireless communication. The wireless communication may use any communication standard or protocol, including but not limited to GSM (Global System for Mobile communications), GPRS (General Packet Radio Service), CDMA2000 (Code Division Multiple Access 2000 ), WCDMA (Wideband Code Division Multiple Access), TD-SCDMA (Time Division-Synchronous Code Division Multiple Access), FDD-LTE (Frequency Division duplex-Long Term Evolution), TDD-LTE (Time Division duplex-Long Term Evolution, time Division Long Term Evolution), 5G, and so on.
WiFi belongs to short-distance wireless transmission technology, and the intelligent terminal can help a user to receive and send e-mails, browse webpages, access streaming media and the like through the WiFi module 102, and provides wireless broadband internet access for the user. Although fig. 1 shows the WiFi module 102, it is understood that it does not belong to the essential constitution of the smart terminal, and may be omitted entirely as needed within the scope not changing the essence of the invention.
The audio output unit 103 may convert audio data received by the radio frequency unit 101 or the WiFi module 102 or stored in the memory 109 into an audio signal and output as sound when the smart terminal 100 is in a call signal reception mode, a call mode, a recording mode, a voice recognition mode, a broadcast reception mode, or the like. Also, the audio output unit 103 may also provide audio output related to a specific function performed by the smart terminal 100 (e.g., a call signal reception sound, a message reception sound, etc.). The audio output unit 103 may include a speaker, a buzzer, and the like.
The a/V input unit 104 is used to receive audio or video signals. The a/V input Unit 104 may include a Graphics Processing Unit (GPU) 1041 and a microphone 1042, and the Graphics processor 1041 processes image data of still pictures or video obtained by an image capturing device (e.g., a camera) in a video capturing mode or an image capturing mode. The processed image frames may be displayed on the display unit 106. The image frames processed by the graphic processor 1041 may be stored in the memory 109 (or other storage medium) or transmitted via the radio frequency unit 101 or the WiFi module 102. The microphone 1042 may receive sounds (audio data) via the microphone 1042 in a phone call mode, a recording mode, a voice recognition mode, or the like, and may be capable of processing such sounds into audio data. The processed audio (voice) data may be converted into a format output transmittable to a mobile communication base station via the radio frequency unit 101 in case of the phone call mode. The microphone 1042 may implement various types of noise cancellation (or suppression) algorithms to cancel (or suppress) noise or interference generated in the course of receiving and transmitting audio signals.
The smart terminal 100 also includes at least one sensor 105, such as a light sensor, a motion sensor, and other sensors. Optionally, the light sensor includes an ambient light sensor and a proximity sensor, the ambient light sensor may adjust the brightness of the display panel 1061 according to the brightness of ambient light, and the proximity sensor may turn off the display panel 1061 and/or the backlight when the smart terminal 100 moves to the ear. As one of the motion sensors, the accelerometer sensor can detect the magnitude of acceleration in each direction (generally, three axes), can detect the magnitude and direction of gravity when stationary, and can be used for applications of recognizing the gesture of the mobile phone (such as horizontal and vertical screen switching, related games, magnetometer gesture calibration), vibration recognition related functions (such as pedometer and tapping), and the like; as for other sensors such as a fingerprint sensor, a pressure sensor, an iris sensor, a molecular sensor, a gyroscope, a barometer, a hygrometer, a thermometer, and an infrared sensor, which can be configured on the mobile phone, the description is omitted here.
The display unit 106 is used to display information input by a user or information provided to the user. The Display unit 106 may include a Display panel 1061, and the Display panel 1061 may be configured in the form of a Liquid Crystal Display (LCD), an Organic Light-Emitting Diode (OLED), or the like.
The user input unit 107 may be used to receive input numeric or character information and generate key signal inputs related to user settings and function control of the intelligent terminal. Alternatively, the user input unit 107 may include a touch panel 1071 and other input devices 1072. The touch panel 1071, also referred to as a touch screen, may collect a touch operation performed by a user on or near the touch panel 1071 (e.g., an operation performed by the user on or near the touch panel 1071 using a finger, a stylus, or any other suitable object or accessory), and drive a corresponding connection device according to a predetermined program. The touch panel 1071 may include two parts of a touch detection device and a touch controller. Optionally, the touch detection device detects a touch orientation of a user, detects a signal caused by a touch operation, and transmits the signal to the touch controller; the touch controller receives touch information from the touch sensing device, converts the touch information into touch point coordinates, sends the touch point coordinates to the processor 110, and can receive and execute commands sent by the processor 110. In addition, the touch panel 1071 may be implemented in various types, such as a resistive type, a capacitive type, an infrared ray, and a surface acoustic wave. The user input unit 107 may include other input devices 1072 in addition to the touch panel 1071. Optionally, other input devices 1072 may include, but are not limited to, one or more of a physical keyboard, function keys (e.g., volume control keys, switch keys, etc.), a trackball, a mouse, a joystick, and the like, and are not limited thereto.
Alternatively, the touch panel 1071 may cover the display panel 1061, and when the touch panel 1071 detects a touch operation thereon or nearby, the touch panel 1071 transmits the touch operation to the processor 110 to determine the type of the touch event, and then the processor 110 provides a corresponding visual output on the display panel 1061 according to the type of the touch event. Although in fig. 1, the touch panel 1071 and the display panel 1061 are two independent components to implement the input and output functions of the intelligent terminal, in some embodiments, the touch panel 1071 and the display panel 1061 may be integrated to implement the input and output functions of the intelligent terminal, which is not limited herein.
The interface unit 108 serves as an interface through which at least one external device is connected to the smart terminal 100. For example, the external device may include a wired or wireless headset port, an external power supply (or battery charger) port, a wired or wireless data port, a memory card port, a port for connecting a device having an identification module, an audio input/output (I/O) port, a video I/O port, an earphone port, and the like. The interface unit 108 may be used to receive input (e.g., data information, power, etc.) from an external device and transmit the received input to one or more elements within the smart terminal 100 or may be used to transmit data between the smart terminal 100 and the external device.
The memory 109 may be used to store software programs as well as various data. The memory 109 may mainly include a program storage area and a data storage area, and optionally, the program storage area may store an operating system, an application program (such as a sound playing function, an image playing function, and the like) required by at least one function, and the like; the storage data area may store data (such as audio data, a phonebook, etc.) created according to the use of the cellular phone, and the like. Further, the memory 109 may include high speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other volatile solid state storage device.
The processor 110 is a control center of the intelligent terminal, connects various parts of the entire intelligent terminal using various interfaces and lines, and performs various functions of the intelligent terminal and processes data by operating or executing software programs and/or modules stored in the memory 109 and calling data stored in the memory 109, thereby performing overall monitoring of the intelligent terminal. Processor 110 may include one or more processing units; preferably, the processor 110 may integrate an application processor and a modem processor, optionally the application processor primarily handles operating systems, user interfaces, application programs, etc., and the modem processor primarily handles wireless communications. It will be appreciated that the modem processor described above may not be integrated into the processor 110.
The intelligent terminal 100 may further include a power supply 111 (such as a battery) for supplying power to various components, and preferably, the power supply 111 may be logically connected to the processor 110 through a power management system, so as to implement functions of managing charging, discharging, and power consumption through the power management system.
Although not shown in fig. 1, the smart terminal 100 may further include a bluetooth module or the like, which is not described herein.
In order to facilitate understanding of the embodiments of the present application, a communication network system on which the intelligent terminal of the present application is based is described below.
Referring to fig. 2, fig. 2 is an architecture diagram of a communication Network system according to an embodiment of the present disclosure, where the communication Network system is an LTE system of a universal mobile telecommunications technology, and the LTE system includes a UE (User Equipment) 201, an e-UTRAN (Evolved UMTS Terrestrial Radio Access Network) 202, an epc (Evolved Packet Core) 203, and an IP service 204 of an operator, which are in communication connection in sequence.
Optionally, the UE201 may be the terminal 100 described above, and is not described herein again.
The E-UTRAN202 includes eNodeB2021 and other eNodeBs 2022, among others. Alternatively, the eNodeB2021 may be connected with other enodebs 2022 through a backhaul (e.g., X2 interface), the eNodeB2021 is connected to the EPC203, and the eNodeB2021 may provide the UE201 access to the EPC 203.
The EPC203 may include an MME (Mobility Management Entity) 2031, an hss (Home Subscriber Server) 2032, other MMEs 2033, an SGW (Serving gateway) 2034, a pgw (PDN gateway) 2035, and a PCRF (Policy and Charging Rules Function) 2036, and the like. Optionally, the MME2031 is a control node that handles signaling between the UE201 and the EPC203, providing bearer and connection management. HSS2032 is used to provide registers to manage functions such as home location register (not shown) and holds subscriber specific information about service characteristics, data rates, etc. All user data may be sent through SGW2034, PGW2035 may provide IP address assignment for UE201 and other functions, and PCRF2036 is a policy and charging control policy decision point for traffic data flow and IP bearer resources, which selects and provides available policy and charging control decisions for a policy and charging enforcement function (not shown).
Although the LTE system is described as an example, it should be understood by those skilled in the art that the present application is not limited to the LTE system, but may also be applied to other wireless communication systems, such as GSM, CDMA2000, WCDMA, TD-SCDMA, 5G, and future new network systems (e.g. 6G), and the like.
Based on the above intelligent terminal hardware structure and communication network system, various embodiments of the present application are provided.
First embodiment
Referring to fig. 3, fig. 3 is a flowchart illustrating a first embodiment of a screen recording method, where the method includes:
s10: responding to a screen recording instruction, and displaying a selection control on a screen recording interface;
s20: receiving a trigger operation aiming at the selection control, and determining or generating an audio processing mode based on the trigger operation;
s30: and when the audio processing mode is noise reduction processing, performing noise reduction processing on the audio data corresponding to the screen recording process according to the target voiceprint.
In the embodiment of the application, the method is applied to an intelligent terminal with a screen recording function, the screen recording instruction is used for triggering the screen recording function, the screen recording function is used for recording screen content to be recorded corresponding to a recorded object, the screen content to be recorded corresponding to the recorded object can be full-screen display content of the intelligent terminal, also can be display content of a display area where at least one display window in a plurality of display windows of the intelligent terminal is located, and also can be display content of at least one running application of a plurality of current running applications of the intelligent terminal, the screen recording function comprises an audio screen recording function and a silent screen recording function, the audio screen recording function is used for executing screen recording in an executing screen recording process and simultaneously recording audio data, audio sources corresponding to the audio data comprise an equipment audio source and an external audio source, the equipment audio source is used for representing the audio data played by the intelligent terminal, the external audio source comprises environmental sound of the environment of the intelligent terminal, and the environmental sound comprises environmental noise, user speaking sound and the like.
Optionally, after receiving the screen recording instruction, responding to the screen recording instruction, and displaying a selection control on a screen recording interface, where the selection control is used to prompt a user to select an audio processing mode, the audio processing mode includes noise reduction processing and non-noise reduction processing, and the audio processing mode is used to indicate that a screen recording function corresponding to the screen recording instruction is an audio screen recording function, and a processing mode of audio data is performed. Optionally, the step S10 further includes:
responding to the screen recording instruction, and determining or generating a screen recording function corresponding to the screen recording instruction;
when the screen recording function corresponding to the screen recording instruction is the sound screen recording function, executing a selection control displayed on a screen recording interface;
optionally, after the step of determining or generating the screen recording function corresponding to the screen recording instruction, the method further includes:
and when the screen recording function is a silent screen recording function, performing silent screen recording operation on the screen content to be recorded corresponding to the recording object so as to determine or generate a screen recording file containing the screen content to be recorded.
Optionally, the manner of determining or generating the screen recording function corresponding to the screen recording instruction may be determined according to a trigger action that triggers the screen recording instruction, where different trigger actions correspond to different screen recording functions, for example, receiving a trigger action of a user on a screen recording control, when the trigger action is a double-click action or a multiple-continuous-click action, the corresponding screen recording function is a sound screen recording function, when the trigger action is a single-click action, the corresponding screen recording function is a non-sound screen recording function, and triggering the corresponding screen recording function based on different trigger actions prevents a microphone option displayed on a screen recording interface from autonomously selecting to click or turn on a microphone after the user turns on the screen recording, so that audio data that does not need to be recorded (invalid audio data) is recorded before the microphone is turned off or audio data that needs to be recorded (valid audio data) before the microphone is turned on, and accuracy of the screen recording is improved.
Optionally, the mode of determining or generating the screen recording function corresponding to the screen recording instruction may also be determined based on setting parameters, where the setting parameters include the screen recording function, and the setting parameters may be a setting before a system leaves a factory, or may also be a setting defined by a user, and this is not limited here.
Optionally, in another embodiment, referring to fig. 4, the step S10 includes:
s11: responding to the screen recording instruction, and outputting a screen recording mode selection list;
s12: and receiving selection operation aiming at the screen recording mode selection control, and displaying the selection control in a screen recording interface.
Optionally, when the screen recording instruction is received, a screen recording mode selection list is output, and the screen recording mode selection list is used for determining the screen recording mode. Optionally, the screen recording modes include a first screen recording mode, a second screen recording mode, a third screen recording mode, a fourth screen recording mode and a fifth screen recording mode, where the first screen recording mode is used to indicate that audio data is not recorded in the screen recording process, the second screen recording mode is used to indicate that internal audio data (sound played by the intelligent terminal) is recorded in the screen recording process and external audio data is not recorded, the third screen recording mode is used to indicate that external audio data is recorded in the screen recording process and internal audio data is not recorded, the fourth screen recording mode is used to indicate that internal audio data and external audio data are recorded in the screen recording process, the fifth screen recording mode is used to indicate that target audio data corresponding to the target voiceprint is recorded in the screen recording process and other audio data except the audio data corresponding to the target voiceprint is not recorded, the target audio data may be audio data in the external audio data that matches the target voiceprint, and the target audio data may be audio data in the internal audio data that matches the target voiceprint. Optionally, when the screen recording mode corresponding to the screen recording instruction is a first screen recording mode, the screen recording function corresponding to the screen recording instruction is a silent screen recording function, and when the screen recording mode corresponding to the screen recording instruction is another screen recording mode except the first screen recording mode, the screen recording function corresponding to the screen recording instruction is a sound screen recording function.
Optionally, the mode of outputting the screen recording mode selection list may be displayed in a superimposed manner in the current display interface of the intelligent terminal, or may be a mode of responding to the screen recording instruction, jumping to a screen recording setting interface, and displaying the screen recording mode selection control on the screen recording setting interface. Optionally, when a selection operation for the screen recording mode selection list is received, the screen recording mode corresponding to the recording instruction is determined based on the selection operation, a screen recording function is determined according to the screen recording mode, and when the screen recording function is determined to be an audio screen recording function, the selection control is displayed on a screen recording interface. Optionally, when it is determined that the screen recording function is a silent screen recording function, performing silent screen recording operation on the screen to be recorded corresponding to the recording object to generate a screen recording file containing the screen to be recorded and not containing audio data.
Optionally, when the screen recording function is a silent screen recording function, starting a screen recording, rejecting recording audio data, and performing a screen recording operation on the screen to be recorded corresponding to the recorded object, and when the recording is finished, generating a screen recording file based on the screen to be recorded corresponding to the recorded object, where the screen recording file is composed of the screen to be recorded corresponding to the recorded object and does not include audio data.
Optionally, when the screen recording function is an audio screen recording function, displaying a selection control on a screen recording interface, where the screen recording interface may be a current display interface of the intelligent terminal, and displaying the selection control in the current display interface in an overlapping manner, and optionally, when the recording object is a non-full-screen display content, displaying the selection control in an overlapping manner in other display areas except for the display area where the recording object is located, so as to prevent the selection control from blocking the screen content to be recorded in the display area where the screen recording object is located.
Optionally, when the screen recording function corresponding to the recording instruction is an audio screen recording function, displaying the selection control on a screen recording interface, receiving a trigger operation for the selection control, and determining or generating an audio processing mode based on the trigger operation, where the audio processing mode includes noise reduction processing and non-noise reduction processing. The noise reduction processing is used for indicating that other audio data except the target audio data corresponding to the target voiceprint is not recorded and only the target audio data is recorded in the screen recording process, namely the noise reduction processing can be used for eliminating other audio data except the target audio data in the internal audio data and can also be used for eliminating other audio data except the target audio data in the external audio data.
Optionally, the method further comprises:
responding to the screen recording instruction, and outputting a screen recording mode selection list;
receiving a selection operation aiming at the screen recording mode selection list, and determining or generating a screen recording mode corresponding to the screen recording instruction according to the selection operation;
when the screen recording mode is the fifth screen recording mode, performing noise reduction processing on audio data corresponding to the screen recording process according to the target voiceprint;
when the screen recording mode is a first screen recording mode, refusing to acquire audio data;
and when the screen recording mode is the second screen recording mode, the third screen recording mode and the fourth screen recording mode, displaying a selection control on a screen recording interface.
Optionally, when it is determined that the screen recording mode is the fifth screen recording mode based on the selection operation for the screen recording mode selection list, it is directly determined that the audio processing mode is the noise reduction processing without displaying a selection control on a screen recording interface.
Optionally, when it is determined that the audio processing mode is noise reduction processing, a target voiceprint is obtained, and noise reduction processing is performed on audio data corresponding to a screen recording process based on the target voiceprint.
Optionally, the target voiceprint may be a preset voiceprint, the target voiceprint may be obtained by outputting a voiceprint selection list when it is determined that the audio processing mode is noise reduction processing, the voiceprint selection list displays a plurality of preset voiceprints, the target voiceprint is determined or generated in response to a selection operation of the voiceprint selection list, optionally, the preset voiceprint in the voiceprint selection list may be that a user triggers a voiceprint adding instruction before triggering a screen recording instruction, the intelligent terminal calls a recording function of the intelligent terminal after receiving the voiceprint adding instruction, displays a recording interface to start recording, determines or generates a to-be-added voiceprint according to a recorded recording file after recording is finished, and adds the to-be-added voiceprint to the preset voiceprint list so that the user can select the preset voiceprint recorded in advance when selecting the target voiceprint, and the preset voiceprint may be a voiceprint of the user himself or may be a voiceprint of another preset user other than the user. Illustratively, when starting recording, a user plays audio data containing sound of a preset user, the intelligent terminal records the played audio data, so as to obtain a voiceprint of the preset user based on the recorded audio data, or uploads the audio data containing the sound of the preset user, and identifies the audio data based on the uploaded audio data, so as to obtain the voiceprint of the preset user, the audio data containing the sound of the preset user may be local audio data that is downloaded by the user to a local storage of the intelligent terminal in advance, or network audio data that is selected by the user in a network, when the audio data containing the sound of the preset user is an online file, the intelligent terminal directly downloads the online file to the local storage, and if the preset voiceprint that the user needs to determine is a leadership voiceprint, after the user uploads the leadership historical speech audio data, the intelligent terminal identifies the obtained voiceprint based on the historical speech data, uses the obtained voiceprint as the preset voiceprint, executes a process of recording and then performs a process of recording a leadership screen, so that the intelligent terminal can record only other audio data without outputting the recorded voiceprint.
Optionally, the target voiceprint may be obtained in real time, and referring to fig. 5, the step S30 includes:
s31: when the audio processing mode is noise reduction processing, displaying a voice input window in a screen recording interface;
s32: receiving a voice signal input aiming at the voice input window, and acquiring the target voiceprint based on the voice signal;
s33: and performing noise reduction processing on the audio data corresponding to the screen recording process according to the target voiceprint.
Optionally, when the screen recording mode corresponding to the screen recording instruction is another screen recording mode except the first screen recording mode, and the audio processing mode is noise reduction processing, a voice input window is displayed in the screen recording interface, where the voice input window is used to prompt a user to input a voice signal for determining a target voiceprint, and the voice signal may be a sound emitted by the user, or may be a voice signal corresponding to audio data including the target voiceprint played by the user.
Optionally, after the voice signal is collected, an acoustic algorithm is adopted to recognize the voice signal to obtain a target voiceprint in the voice signal, and optionally, when the voice signal is recognized to correspond to at least two voiceprints, a voiceprint selection list is output based on the recognized voiceprints, so that a user selects at least one voiceprint based on the voiceprint selection list, and the at least one voiceprint is determined as the target voiceprint.
Optionally, the target voiceprint may also be determined based on a recording object, obtain audio data corresponding to the recording object, identify, by using an acoustic algorithm, the audio data corresponding to the recording object to identify at least one voiceprint corresponding to the recording object, and use the at least one voiceprint corresponding to the recording object as the target voiceprint. Optionally, when the recording object is an application, at least one voiceprint corresponding to the application is obtained according to the audio data corresponding to the sound source of the application, so that when the recording object is actually recorded, only the audio data corresponding to the recording object is recorded through the at least one voiceprint corresponding to the recording object, and the audio data of other recording objects except the recording object is not recorded, thereby improving user experience.
Optionally, when the audio processing mode is noise reduction processing, noise reduction processing is performed on audio data corresponding to a screen recording process according to a target voiceprint, the audio data corresponding to the screen recording process is determined according to a screen recording mode corresponding to the screen recording instruction, when the screen recording mode is a second screen recording mode, the audio data corresponding to the screen recording process is internal audio data, when the screen recording mode is a third screen recording mode, the audio data corresponding to the screen recording process is external audio data, when the screen recording mode is a fourth screen recording mode, the audio data corresponding to the screen recording process is internal audio data and external audio data, and when the screen recording mode is a fifth screen recording mode, the audio data corresponding to the screen recording process is internal audio data and/or external audio data.
Optionally, the audio data corresponding to the screen recording process is subjected to noise reduction processing according to the target voiceprint by setting the audio data of the recording threads of other voiceprints except the target voiceprint to 0 when the screen display content of the recording object is recorded. In this way, only the target audio data corresponding to the recording object and matching the target voiceprint can be recorded, and other audio data corresponding to the recording object and not matching the target voiceprint can not be recorded. Optionally, when the screen recording mode corresponding to the screen recording instruction is a second screen recording mode, recording target internal audio data matched with the target voiceprint in the internal audio data according to the target voiceprint, and rejecting recording other content audio data unmatched with the target voiceprint in the internal audio data, so as to implement noise reduction processing on the internal audio data; when the screen recording mode corresponding to the screen recording instruction is a third screen recording mode, recording target external audio data matched with the target voiceprint in the external audio data according to the target voiceprint, and refusing to record other external audio data unmatched with the target voiceprint in the external audio data so as to realize noise reduction processing on the external audio data; when the screen recording mode corresponding to the screen recording instruction is a fourth screen recording mode, recording target internal audio data matched with the target voiceprint in internal audio data and target external audio data matched with the target voiceprint in external audio data according to the target voiceprint so as to perform noise reduction on the internal audio data and the external audio data at the same time; when the screen recording mode is a fifth screen recording mode, when the audio data corresponding to the target voiceprint is internal audio data, recording target internal audio data matched with the target voiceprint in the internal audio data, and/or when the audio data corresponding to the target voiceprint is external audio data, recording target external audio data matched with the target voiceprint in the external audio data so as to record only the target audio data corresponding to the target voiceprint.
Optionally, when the screen recording function corresponding to the screen recording instruction is a sound screen recording function and a screen recording stopping instruction for the screen recording process is received, generating a recording file based on the screen content to be recorded corresponding to the recording object and the target audio data corresponding to the target voiceprint; and when the screen recording function corresponding to the screen recording instruction is a silent screen recording function, generating a recording file based on the screen content to be recorded corresponding to the recording object when a screen recording stopping instruction aiming at the screen recording process is received.
In the embodiment of the application, a recording instruction is responded, a screen recording mode selection list is output so that a user can select a screen recording mode, whether a screen recording function corresponding to the screen recording instruction is a sound screen recording function is determined based on a selection result corresponding to the screen recording mode, a selection control is displayed on a screen recording interface when the screen recording function corresponding to the recording instruction is determined to be the sound screen recording function, the selection control is used for determining or generating an audio processing mode, a target voiceprint is obtained when the audio processing mode is determined or generated to be noise reduction processing, noise reduction processing is performed on audio data corresponding to a screen recording process based on the target voiceprint, so that other audio data except the target audio data corresponding to the target voiceprint are not recorded in the recording process, the audio data in a recording file do not contain other audio data except the target audio data corresponding to the target voiceprint, the effects of not recording invalid audio data and only recording valid audio data are achieved, and user experience is improved.
Second embodiment
Based on the first embodiment, referring to fig. 6, after the step S20, the method further includes:
s21, after a screen recording process is started based on the audio processing mode, displaying an audio recording control in the screen recording interface;
and S22, receiving a first operation aiming at the audio recording control, and interrupting recording of audio data.
Optionally, when it is determined that the screen recording function corresponding to the screen recording instruction is an audio screen recording function, after a screen recording process is started based on the audio processing mode, an audio recording control is displayed on the screen recording interface, and the audio recording control is used for controlling the audio data to be recorded in an interrupted manner and stopping recording. Optionally, when the operation on the audio recording control is a first operation, interrupting recording of audio data; and when the operation aiming at the audio recording control is a second operation, stopping recording the audio data, wherein the screen recording mode is a first screen recording mode, namely only recording the screen content to be recorded, and not recording the audio data.
Optionally, when a first operation for the audio recording control is received, the recording of the audio data is interrupted, and before the interruption of the recording of the recorded audio data is interrupted, the intelligent terminal determines the target audio data based on an audio processing mode and a screen recording mode. Optionally, when the audio processing mode is a noise reduction processing mode, when a screen recording mode corresponding to the screen recording instruction is a second screen recording mode, recording of target internal audio data matched with the target voiceprint in the internal audio data is interrupted, when the screen recording mode is a third screen recording mode, recording of target external audio data matched with the target voiceprint in the external audio data is interrupted, when the screen recording mode is a fourth screen recording mode, recording of target internal audio data matched with the target voiceprint in the internal audio data and recording of target external audio data matched with the target voiceprint in the external audio data are interrupted, and when the screen recording mode is a fifth screen recording mode, recording of target audio data corresponding to the target voiceprint is interrupted.
Optionally, when the audio processing mode is a non-noise reduction processing mode, when the screen recording mode corresponding to the screen recording instruction is a second screen recording mode, the recording of the internal audio data is interrupted, when the screen recording mode is a third screen recording mode, the recording of the external audio data is interrupted, when the screen recording mode is a fourth screen recording mode, the recording of the internal audio data and the recording of the external audio data are interrupted, and when the screen recording mode is a fifth screen recording mode, the recording of the target audio data corresponding to the target voiceprint is stopped.
Optionally, referring to fig. 7, after the step S22, the method further includes:
s23: and receiving the starting operation aiming at the audio recording control, and continuously recording the audio data.
Optionally, the audio recording control is further configured to control audio data to start recording, and after the recording of the audio data is interrupted, when a start operation for the audio recording control is received, the audio data is continuously recorded, and optionally, the continuously recorded audio data corresponds to the audio processing mode, that is, the continuously recorded audio data is audio data corresponding to a screen recording process, and target audio data matched with the target voiceprint is generated after the noise reduction processing is performed on the audio data.
Optionally, in order to ensure consistency between the time point of the content to be recorded corresponding to the recording object and the time point of the audio data, audio data between the interruption time point corresponding to the interruption of recording the audio data and the duration time point of the duration of recording the audio data is set as blank audio data.
In the embodiment of the application, the audio recording control is displayed on the screen recording interface in the screen recording process, so that when a user needs to record a recording file which does not contain audio data, the recording of the audio data is interrupted or the recording of the audio data is stopped, and when the user needs to record the recording file which contains the audio data, the recording of the audio data is continued, and the flexibility of recording the audio data is improved.
Third embodiment
Based on the above embodiment, referring to fig. 8, after the step S20, the method further includes:
s40: when the audio processing mode is non-noise reduction processing, acquiring volume values of other audio data except the target audio data corresponding to the target voiceprint;
s41: and when the volume value is greater than or equal to the preset volume value, displaying a selection control on a recording screen interface and/or displaying an audio recording control.
Optionally, when the audio processing mode is non-noise reduction processing, the screen recording mode corresponding to the screen recording instruction includes a second screen recording mode, a third screen recording mode and a fourth screen recording mode, and when the screen recording mode is the second screen recording mode, internal audio data are acquired in the screen recording process; when the screen recording mode is the third screen recording mode, acquiring external audio data in the screen recording process; and when the screen recording mode is the fourth screen recording mode, acquiring internal audio data and external audio data in the screen recording process.
Optionally, in the process of recording the screen, when the audio processing mode is non-noise reduction processing, the recorded audio file is likely to contain invalid audio data, and in order to facilitate a user to adjust the audio processing mode when the invalid audio data appears, so as to change the audio processing mode from the non-noise reduction processing to noise reduction processing, or enable the user to interrupt recording of the invalid audio data, in the embodiment of the application, whether the audio data corresponding to the screen recording process meets a preset condition is detected in real time in the process of executing the screen recording based on the non-noise reduction processing, and if so, a selection control is displayed on a screen recording interface for the user to update the audio processing mode, and/or an audio recording control is displayed for the user to trigger a stop operation for the audio recording control, so as to interrupt or stop recording the audio data. Optionally, when it is detected that the audio data meets the preset condition, a screen recording mode selection list may be further output, so that the user may select to continue to perform the screen recording operation in the first screen recording mode or continue to perform the screen recording operation in the second screen recording mode.
Optionally, the preset condition may be that the volume value of other audio data except the target audio data corresponding to the target voiceprint is greater than or equal to a preset volume value, for example, when the screen recording mode is the second screen recording mode, the volume value of other audio data in the internal audio data is higher, or when the screen recording mode is the third screen recording mode, the volume value of external noise is larger, or when the screen recording mode is the fourth screen recording mode, the volume value of other audio data in the internal audio data is higher and/or the volume value of external noise is larger, it is determined that the audio data corresponding to the screen recording process meets the preset condition.
Optionally, the preset condition may also be that the frequency of occurrence of the other audio data is greater than or equal to a preset frequency, for example, the frequency of occurrence of noise is high, and the frequency of occurrence of the audio data of other users is high, which means that the recording file contains a large amount of invalid audio data, which easily causes the invalid audio data in the recording file to occupy a large amount of storage space.
Optionally, when the preset condition further includes that the audio data corresponding to the screen recording process does not include the target audio data within a preset time period, it is determined that the preset condition is met, so as to reduce a storage space required to be consumed by the audio data.
In the embodiment of the application, when the audio processing mode is determined to be non-noise reduction processing, whether the audio data corresponding to the screen recording process meets the preset condition or not is determined by obtaining the volume value of other audio data and/or the occurrence frequency of other audio data and/or the judgment result of whether the audio data corresponding to the screen recording process includes the target audio data within the preset time, and when the preset condition is determined to be met, the selection control is displayed on the screen recording interface and/or the audio recording control is displayed, so that the situation that the recording file includes a large amount of invalid audio data and the invalid audio data shields the valid audio data is avoided, and in the screen recording process, the switching efficiency of the user for switching the audio processing mode is improved, the switching efficiency of the screen recording mode is improved, and the flexibility of audio data recording is improved.
Fourth embodiment
Based on the above embodiment, referring to fig. 9, after the step S30, the method further includes:
s50: receiving a screen recording stopping instruction aiming at the screen recording process, and determining or generating a recording file according to the processed audio data and the video data corresponding to the screen recording process;
s60: determining or generating an audio time axis according to the audio data, wherein the audio time axis is used for indicating a time interval containing the audio data and a time interval not containing the audio data;
s70: and playing the recording file in a preview interface, and displaying the audio time axis in a superposition manner on the preview interface.
Optionally, when a screen recording stopping instruction for the screen recording process is received, the screen recording interface displays a screen recording control for controlling the audio recording interruption, the continuous recording and the stopping of the recording by the user, and also displays a screen recording control for triggering the screen recording stopping and the screen recording interruption. For example, referring to fig. 10, fig. 10 shows a schematic view of a screen recording interface, optionally, different trigger actions for a screen recording control correspond to different instructions, if the trigger action is a single-click action, it is determined to trigger an instruction to stop screen recording for the screen recording process, if the trigger action is a double-click action, it is determined to trigger an instruction to stop screen recording for the screen recording process, after it is determined to trigger the instruction to stop screen recording for the screen recording process, when the single-click action is received, it is determined to trigger an instruction to start screen recording for the screen recording process to continue screen recording, and when the double-click action is received, it is determined to trigger an instruction to stop screen recording for the screen recording process. Optionally, the manner of triggering the screen recording stopping instruction may also be triggered according to a gesture action, such as sliding down along an upper edge of the intelligent terminal or sliding up along a lower edge of the intelligent terminal, or according to a voice instruction input by a user, which is not limited herein.
Optionally, after receiving a screen recording stopping instruction, determining or generating a recording file from the processed audio data and the video data in the screen recording process, where the processed audio data is target audio data corresponding to the target voiceprint, and the video data is screen recording generation of screen to be recorded corresponding to a recording object.
Optionally, after the recording file is determined or generated, a preview control may be output, where the preview control is used for a user to confirm whether to preview the recording file. Optionally, when a trigger operation for the preview control is received, an audio time axis is determined or generated according to the audio data, where the audio time axis includes multiple time points, the time points are in one-to-one correspondence with the video time points corresponding to the video data, the audio time axis includes a time interval for indicating that the audio data is included and a time interval that does not include the audio data, that is, a recording file corresponding to the time interval that includes the audio data includes both the video data and the audio data, and a recording file corresponding to the time interval that does not include the audio data only includes the video data but not the audio data.
Optionally, after the audio time axis is determined or generated, playing the recording file in a preview interface, where the playing mode of the recording file is to play the video data, and simultaneously playing the audio data corresponding to the current playing time point according to the current playing time point corresponding to the video data, so as to synchronously play the audio data and the video data.
Optionally, in the process of playing the recording file, the audio timeline is displayed in the preview interface in an overlapping manner, so that the audio timeline is displayed in the currently played video data in an overlapping manner. Optionally, different types of time intervals in the audio time axis are displayed in different display manners, where the display manners include display colors, and if a recording file corresponding to a time interval contains audio data, the audio time axis corresponding to the time interval is displayed in red, and if the recording file corresponding to the time interval does not contain audio data, the audio time axis corresponding to the time interval is displayed in yellow; the display mode further comprises a display size, wherein the display size comprises the size of an audio time axis corresponding to a time interval in the length direction and/or the size of an audio time axis corresponding to a time period in the width mode, if a recording file corresponding to the time interval contains audio data, the distance of each progress point in the audio time axis corresponding to the time interval in the length mode is enlarged to enlarge the size of the audio time axis corresponding to the time interval in the length direction, and if the recording file corresponding to the time interval does not contain the audio data, the distance of each progress point in the audio time axis corresponding to the time interval in the length mode is reduced to reduce the size of the audio time axis corresponding to the time interval in the length direction; the display mode further comprises display identifiers, different types of time intervals correspond to different display identifiers, if the recording files corresponding to the time intervals contain audio data, the display identifiers corresponding to the time intervals are 'sound', and if the recording files corresponding to the time intervals do not contain audio data, the display identifiers corresponding to the time intervals are 'mute'. Exemplarily, referring to fig. 11, fig. 11 shows a schematic diagram of an audio time axis.
Optionally, in the process of playing the recording file on a preview interface, the audio time axis is displayed in a superimposed manner on the preview interface, and a user knows which recording files corresponding to which time periods contain audio data and which recording files corresponding to which time periods do not contain audio data based on the audio time axis.
Optionally, referring to fig. 12, the step of playing the recording file in the preview interface further includes:
s71: acquiring a current playing time point in the playing process of video data, or acquiring a selected time point aiming at the audio time axis;
s72: taking the current playing time point or the selected time point as a preview time point;
s73: acquiring audio data matched with the preview time point, and determining or generating target information according to the audio data;
s74: and superposing the target information on a target video frame corresponding to the preview time point, and displaying the superposed target image frame.
Optionally, when the audio data in the recording file is played in a sound playing manner, the audio data may be converted into corresponding target information and displayed in the currently played video data.
Optionally, a preview time point is obtained, the audio data corresponding to the preview time point is audio data to be played, the preview time point is obtained in a manner of obtaining a current playing time point in a playing process of the video data, based on synchronization of the time point of the video data and the time point of the audio data, the audio data corresponding to the current playing time point can be determined as the audio data to be played according to the current playing time point, optionally, the preview time point is obtained in a manner of being selected by a user, a trigger action of the user for an audio time axis is received, a time point corresponding to a trigger position corresponding to the trigger action is taken as a selected time point selected by the user, the selected time point is taken as the preview time point, and optionally, the selected time point is taken as a time point in a time interval containing the audio data.
Optionally, after the preview time point is obtained, audio data matched with the preview time point is obtained, and target information is determined or generated according to the audio data, where the target information may be text information obtained by converting the audio data through a speech-to-text technique, or text information obtained by converting key audio data in the audio data through a speech-to-text technique.
Optionally, the target information may also be operation description information generated by a first input of the to-be-recorded screen content corresponding to the video data, for example, when a user unlocks for switching between the to-be-recorded screen contents by inputting target audio data, for example, "jump by clicking a confirmation key", the operation description information may be text information such as "click confirmation key", and the operation description information may also be a dynamic image such as a click dynamic image.
Optionally, after the target information is determined or generated, the target information is displayed in a superimposed manner in the target video frame corresponding to the preview time point, so as to "play" the audio data in a display manner. Optionally, the target information may be displayed in a superimposed manner by superimposing the target information as subtitles on one side of the target video frame, or by determining a video frame display position associated with the target information, and displaying the target information in the superimposed manner on the video frame display position or in a preset range of the video frame display position, such as on the left side, the right side, and the like of the video frame display position. Illustratively, the target information is a "click confirmation key", the display position of the associated video frame is a display position of the target video frame displaying the "confirmation key", and the target information is displayed in a superimposed manner on the display position side of the "confirmation key".
Optionally, the target information may be displayed in a manner that the preview interface is divided into at least two display areas, where a first display area is used for displaying the target video frame and a second display area is used for displaying the target information.
In the embodiment of the application, after determining or generating the recording file, through outputting the preview control, for the user to preview the recording file through triggering the preview control, when receiving the trigger operation of the preview control, according to the audio data determine or generate the audio data axis, play in the preview interface the recording file, and will the audio time axis is superposedly displayed in the preview interface, so that the user knows that which time periods correspond to the recording file contain the audio data, which time periods correspond to the recording file do not contain the audio data, play the recording file in-process, through obtaining the preview time point, will with the audio data that the preview time point matches convert into the target information, and then show the target information in the preview interface, realized with the mode "play" audio data that shows, improved the visualization of audio data, the user learns the content of the target audio data more directly perceived based on the target information that shows, promotes user experience.
The application also provides an intelligent terminal, which comprises a memory and a processor, wherein the memory is stored with a screen recording program, and the screen recording program is executed by the processor to realize the steps of the screen recording method in any embodiment.
The application further provides a storage medium, where a screen recording program is stored on the storage medium, and the screen recording program, when executed by the processor, implements the steps of the screen recording method in any of the embodiments.
In the embodiments of the intelligent terminal and the storage medium provided in the present application, all technical features of any one of the embodiments of the screen recording method may be included, and the expanding and explaining contents of the specification are substantially the same as those of each embodiment of the method, and are not described herein again.
Embodiments of the present application also provide a computer program product, which includes computer program code, when the computer program code runs on a computer, the computer is caused to execute the method in the above various possible embodiments.
Embodiments of the present application further provide a chip, which includes a memory and a processor, where the memory is used to store a computer program, and the processor is used to call and run the computer program from the memory, so that a device in which the chip is installed executes the method in the above various possible embodiments.
It is to be understood that the foregoing scenarios are only examples, and do not constitute a limitation on application scenarios of the technical solutions provided in the embodiments of the present application, and the technical solutions of the present application may also be applied to other scenarios. For example, as can be known by those skilled in the art, with the evolution of system architecture and the emergence of new service scenarios, the technical solution provided in the embodiments of the present application is also applicable to similar technical problems.
The above-mentioned serial numbers of the embodiments of the present application are merely for description and do not represent the merits of the embodiments.
The steps in the method of the embodiment of the application can be sequentially adjusted, combined and deleted according to actual needs.
The units in the device of the embodiment of the application can be combined, divided and deleted according to actual needs.
In the present application, the same or similar term concepts, technical solutions and/or application scenario descriptions will be generally described only in detail at the first occurrence, and when the description is repeated later, the detailed description will not be repeated in general for brevity, and when understanding the technical solutions and the like of the present application, reference may be made to the related detailed description before the description for the same or similar term concepts, technical solutions and/or application scenario descriptions and the like which are not described in detail later.
In the present application, each embodiment is described with an emphasis on the description, and reference may be made to the description of other embodiments for parts that are not described or recited in any embodiment.
All possible combinations of the technical features in the embodiments are not described in the present application for the sake of brevity, but should be considered as the scope of the present application as long as there is no contradiction between the combinations of the technical features.
Through the above description of the embodiments, those skilled in the art will clearly understand that the method of the above embodiments can be implemented by software plus a necessary general hardware platform, and certainly can also be implemented by hardware, but in many cases, the former is a better implementation manner. Based on such understanding, the technical solutions of the present application may be embodied in the form of a software product, which is stored in a storage medium (e.g., ROM/RAM, magnetic disk, optical disk) and includes instructions for enabling a terminal device (e.g., a mobile phone, a computer, a server, a controlled terminal, or a network device) to execute the method of each embodiment of the present application.
In the above embodiments, all or part of the implementation may be realized by software, hardware, firmware, or any combination thereof. When implemented in software, may be implemented in whole or in part in the form of a computer program product. The computer program product includes one or more computer instructions. The procedures or functions according to the embodiments of the present application are all or partially generated when the computer program instructions are loaded and executed on a computer. The computer may be a general purpose computer, a special purpose computer, a network of computers, or other programmable device. The computer instructions may be stored on a storage medium or transmitted from one storage medium to another, for example, the computer instructions may be transmitted from one website, computer, server, or data center to another website, computer, server, or data center via wire (e.g., coaxial cable, fiber optic, digital subscriber line) or wirelessly (e.g., infrared, wireless, microwave, etc.). The storage medium may be any available medium that can be accessed by a computer or a data storage device including one or more available media integrated servers, data centers, and the like. The usable medium may be a magnetic medium (e.g., floppy Disk, storage Disk, magnetic tape), an optical medium (e.g., DVD), or a semiconductor medium (e.g., solid State Disk (SSD)), among others.
The above description is only a preferred embodiment of the present application, and not intended to limit the scope of the present application, and all the equivalent structures or equivalent processes that can be directly or indirectly applied to other related technical fields by using the contents of the specification and the drawings of the present application are also included in the scope of the present application.
Claims (10)
1. A screen recording method, characterized in that the method comprises the steps of:
s10: responding to a screen recording instruction, and displaying a selection control on a screen recording interface;
s20: receiving a trigger operation aiming at the selection control, and determining or generating an audio processing mode based on the trigger operation;
s30: and when the audio processing mode is noise reduction processing, performing noise reduction processing on the audio data corresponding to the screen recording process according to the target voiceprint.
2. The method of claim 1, wherein the S30 step comprises:
when the audio processing mode is noise reduction processing, displaying a voice input window in a screen recording interface;
receiving a voice signal input aiming at the voice input window, and acquiring the target voiceprint based on the voice signal;
and performing noise reduction processing on the audio data corresponding to the screen recording process according to the target voiceprint.
3. The method of claim 1 or 2, wherein the step of determining the audio processing mode based on the triggering operation is followed by further comprising:
s21: displaying an audio recording control in the screen recording interface after starting a screen recording process based on the audio processing mode;
s22: and receiving a first operation aiming at the audio recording control, and interrupting recording of audio data.
4. The method of claim 3, wherein after the step of S22, further comprising:
and receiving the starting operation aiming at the audio recording control, and continuously recording the audio data.
5. The method of claim 3, wherein after the step of S20, further comprising:
when the audio processing mode is non-noise reduction processing, acquiring volume values of other audio data except the target audio data corresponding to the target voiceprint;
and when the volume value is greater than or equal to the preset volume value, displaying a selection control on a screen recording interface and/or displaying an audio recording control.
6. The method according to claim 1 or 2, wherein the S10 step comprises:
responding to the screen recording instruction, and outputting a screen recording mode selection list;
and receiving a selection operation aiming at the screen recording mode selection list, and displaying the selection control on a screen recording interface.
7. The method according to claim 1 or 2, wherein after the step S30, further comprising:
receiving a screen recording stopping instruction aiming at the screen recording process, and determining or generating a recording file according to the processed audio data and the video data corresponding to the screen recording process;
determining or generating an audio timeline from the audio data;
and playing the recording file in a preview interface, and displaying the audio time axis in an overlapping manner on the preview interface.
8. The method of claim 7, wherein the step of playing the recording file in the preview interface further comprises:
acquiring a current playing time point in the playing process of video data, or acquiring a selected time point aiming at the audio time axis;
taking the current playing time point or the selected time point as a preview time point;
acquiring audio data matched with the preview time point, and determining or generating target information according to the audio data;
and overlaying the target information in a target video frame corresponding to the preview time point, and displaying the overlaid target image frame.
9. An intelligent terminal, characterized in that, intelligent terminal includes: memory, processor, wherein the memory has stored thereon a screen recording program which, when executed by the processor, implements the steps of the screen recording method according to any one of claims 1 to 8.
10. A storage medium, characterized in that the storage medium has stored thereon a computer program which, when being executed by a processor, carries out the steps of the screen recording method according to any one of claims 1 to 8.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211570405.XA CN115955529A (en) | 2022-12-08 | 2022-12-08 | Screen recording method, intelligent terminal and storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211570405.XA CN115955529A (en) | 2022-12-08 | 2022-12-08 | Screen recording method, intelligent terminal and storage medium |
Publications (1)
Publication Number | Publication Date |
---|---|
CN115955529A true CN115955529A (en) | 2023-04-11 |
Family
ID=87297463
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202211570405.XA Pending CN115955529A (en) | 2022-12-08 | 2022-12-08 | Screen recording method, intelligent terminal and storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN115955529A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116567338A (en) * | 2023-04-14 | 2023-08-08 | 深圳支点电子智能科技有限公司 | Intelligent screen recording method and related device in video conference scene |
-
2022
- 2022-12-08 CN CN202211570405.XA patent/CN115955529A/en active Pending
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116567338A (en) * | 2023-04-14 | 2023-08-08 | 深圳支点电子智能科技有限公司 | Intelligent screen recording method and related device in video conference scene |
CN116567338B (en) * | 2023-04-14 | 2024-01-19 | 深圳支点电子智能科技有限公司 | Intelligent screen recording method and related device in video conference scene |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107229402B (en) | Dynamic screen capturing method and device of terminal and readable storage medium | |
CN110187808B (en) | Dynamic wallpaper setting method and device and computer-readable storage medium | |
CN110855921B (en) | Video recording control method and electronic equipment | |
CN109828711A (en) | A kind of reading management method, mobile terminal and the storage medium of mobile terminal | |
CN113805837A (en) | Audio processing method, mobile terminal and storage medium | |
CN108712558A (en) | Incoming call reminding method, mobile terminal and the readable storage medium storing program for executing of mobile terminal | |
CN115955529A (en) | Screen recording method, intelligent terminal and storage medium | |
CN115914719A (en) | Screen projection display method, intelligent terminal and storage medium | |
CN113867586A (en) | Icon display method, intelligent terminal and storage medium | |
CN113885752A (en) | Icon processing method, intelligent terminal and storage medium | |
CN113900556A (en) | Icon processing method, intelligent terminal and storage medium | |
CN113419694A (en) | Processing method, mobile terminal and storage medium | |
CN113867588A (en) | Icon processing method, intelligent terminal and storage medium | |
CN109669594B (en) | Interaction control method, equipment and computer readable storage medium | |
CN116225601A (en) | Display method, intelligent terminal and storage medium | |
CN115118812A (en) | Information processing method, intelligent terminal and storage medium | |
CN114089897A (en) | Wallpaper switching method, intelligent terminal and storage medium | |
CN115914697A (en) | Screen projection control method, intelligent terminal and storage medium | |
CN114116104A (en) | Information control method, intelligent terminal and storage medium | |
CN115729659A (en) | Application running method, mobile terminal and storage medium | |
CN115793922A (en) | Display method, intelligent terminal and storage medium | |
CN115834758A (en) | Screen recording method, intelligent terminal and storage medium | |
CN115065692A (en) | Multimedia information sharing method, intelligent terminal and storage medium | |
CN115065744A (en) | Configuration method, intelligent terminal and storage medium | |
CN115225757A (en) | Object control method, intelligent terminal and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication |