CN109309762B - Message processing method, device, mobile terminal and storage medium - Google Patents

Message processing method, device, mobile terminal and storage medium Download PDF

Info

Publication number
CN109309762B
CN109309762B CN201811462250.1A CN201811462250A CN109309762B CN 109309762 B CN109309762 B CN 109309762B CN 201811462250 A CN201811462250 A CN 201811462250A CN 109309762 B CN109309762 B CN 109309762B
Authority
CN
China
Prior art keywords
voice
image
target
playing
target area
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201811462250.1A
Other languages
Chinese (zh)
Other versions
CN109309762A (en
Inventor
罗彪
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nubia Technology Co Ltd
Original Assignee
Nubia Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nubia Technology Co Ltd filed Critical Nubia Technology Co Ltd
Priority to CN201811462250.1A priority Critical patent/CN109309762B/en
Publication of CN109309762A publication Critical patent/CN109309762A/en
Application granted granted Critical
Publication of CN109309762B publication Critical patent/CN109309762B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72433User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for voice messaging, e.g. dictaphones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72439User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for image or video messaging

Abstract

The invention discloses a message processing method and device, a mobile terminal and a storage medium, wherein the method comprises the following steps: receiving a message sent by a terminal, and acquiring an image and voice in the message; detecting the type of an operation instruction of a user; if the operation instruction is an image operation instruction, acquiring a target image in each image, and amplifying the target image to acquire a target image interface, wherein different voice playing interfaces are respectively arranged in each preset area of the target image interface; and when the preset area is detected to have a target area which is being amplified, playing a first voice corresponding to the target area in the voices through a voice playing interface of the target area. The invention solves the technical problem that the viewing time sequence of voice and pictures can not be flexibly coordinated in the prior art.

Description

Message processing method, device, mobile terminal and storage medium
Technical Field
The present invention relates to the field of communications technologies, and in particular, to a message processing method and apparatus, a mobile terminal, and a storage medium.
Background
At present, with the rapid development of communication technology, mobile terminals such as mobile phones and tablet computers are getting closer to our lives, and applications of various social software (QQ, wechat, etc.) installed on the mobile terminals are slowly becoming a part of people's daily lives. However, when the user uses the social software, receives pictures and voices sent by other users, and needs to view the pictures first, the pictures need to be enlarged to view the pictures first, and then the pictures are reduced to play the voices. For example, when a user receives an engineering drawing and a corresponding explaining voice sent by other users on a mobile terminal, when the user looks over the engineering drawing and finds a fuzzy part, the user needs to quit the picture and click the voice for confusion, so that the working efficiency of the user is greatly influenced. Therefore, how to flexibly coordinate the viewing timing of the voice and the picture becomes a technical problem to be solved urgently at present.
Disclosure of Invention
The invention mainly aims to provide a message processing method, a message processing device, a mobile terminal and a storage medium, and aims to solve the technical problem that the viewing time sequence of voice and pictures cannot be flexibly coordinated in the prior art.
In order to achieve the above object, the present invention provides a message processing method, which includes the following steps:
receiving a message sent by a terminal, and acquiring an image and voice in the message;
detecting the type of an operation instruction of a user;
if the operation instruction is an image operation instruction, acquiring a target image in each image, and amplifying the target image to acquire a target image interface, wherein different voice playing interfaces are respectively arranged in each preset area of the target image interface;
and when the preset area is detected to have a target area which is being amplified, playing a first voice corresponding to the target area in the voices through a voice playing interface of the target area.
Optionally, the step of playing, when it is detected that a target area which is being amplified exists in the preset area, a first voice corresponding to the target area in the voices through a voice playing interface of the target area includes:
when detecting that a target area which is being amplified exists in the preset area, judging whether the target area has a limiting condition;
and if the limitation condition does not exist, playing a first voice corresponding to the target area in the voices through a voice playing interface of the target area.
Optionally, after the step of determining whether the target area has the limitation condition, the method includes:
if the limiting condition exists, acquiring account information of the user, and judging whether the account information meets the limiting condition;
if the account information meets the limiting condition, playing a first voice corresponding to the target area in the voice through a voice playing interface of the target area;
and if the account information does not meet the limiting conditions, automatically shielding the voice playing interface of the target area.
Optionally, after the step of playing the first voice in the voices through the voice playing interface of the target area, the method includes:
and when the amplification of another preset area in the target image interface relative to the target area is detected, automatically stopping the current first voice playing, and playing a second voice corresponding to the other preset area through a voice playing interface of the other preset area.
Optionally, after the step of respectively setting different voice playing interfaces in each preset area of the target image interface, the method includes:
the voice type of each of the voices is acquired,
acquiring the region type of each preset region, and matching each voice type with each region type to acquire the play-supporting voice type of each preset region;
and determining a target area in each preset area, acquiring a first voice type supported by the target area to be played, and taking the first voice type as a first voice to be played.
Optionally, after the step of detecting the type of the operation instruction of the first user, the method includes:
if the operation instruction is a voice opening instruction, opening target voice in the voices based on the voice opening instruction;
detecting whether voice information triggering the sub-images in each image exists in the target voice;
and if the target voice has the voice information triggering the subimage, automatically starting and amplifying the subimage.
Optionally, if the operation instruction is a voice start instruction, after the step of starting the target voice in the voices based on the voice start instruction, the method includes:
acquiring image information of each image;
and determining voice information triggering the sub-images in each image according to each image information.
Further, to achieve the above object, the present invention provides a message processing apparatus, including:
the receiving module is used for receiving a message sent by a terminal and acquiring an image and voice in the message;
the acquisition module is used for detecting the type of an operation instruction of a user;
the device comprises a setting module, a processing module and a display module, wherein the setting module is used for acquiring a target image in each image and amplifying the target image to acquire a target image interface if the operation instruction is an image operation instruction, and different voice playing interfaces are respectively arranged in each preset area of the target image interface;
and the playing module is used for playing a first voice corresponding to the target area in the voices through a voice playing interface of the target area when detecting that the target area which is being amplified exists in each preset area.
In addition, in order to achieve the above object, the present invention also provides a mobile terminal;
the mobile terminal includes: a memory, a processor, and a computer program stored on the memory and executable on the processor, wherein:
which when executed by the processor implements the steps of the message processing method as described above.
In addition, to achieve the above object, the present invention also provides a storage medium;
the storage medium has stored thereon a computer program which, when executed by a processor, implements the steps of the message processing method as described above.
The invention receives the message sent by the terminal and obtains the image and the voice in the message; detecting the type of an operation instruction of a user; if the operation instruction is an image operation instruction, acquiring a target image in each image, and amplifying the target image to acquire a target image interface, wherein different voice playing interfaces are respectively arranged in each preset area of the target image interface; and when the preset area is detected to have a target area which is being amplified, playing a first voice corresponding to the target area in the voices through a first voice playing interface of the target area. When the target area is enlarged in the target image interface, the first voice corresponding to the target area is automatically played, and the user does not need to manually start the corresponding voice, so that the working efficiency is improved for the user, the intelligent effect of the user is also improved, the user can listen to the corresponding voice when checking the image in the received message, the understanding of the user to the image is accelerated, the use experience of the user is improved, the technical effect that the user can automatically play the corresponding voice when checking the image is achieved, and the technical problem that the checking time sequence of the voice and the checking time sequence of the image cannot be flexibly coordinated in the prior art is also solved.
Drawings
Fig. 1 is a schematic diagram of a hardware structure of an alternative mobile terminal according to an embodiment of the present invention;
FIG. 2 is a diagram of a wireless communication device of the mobile terminal of FIG. 1;
FIG. 3 is a flowchart illustrating a first embodiment of a message processing method according to the present invention;
FIG. 4 is a flowchart illustrating a second embodiment of a message processing method according to the present invention;
FIG. 5 is a functional block diagram of a message processing apparatus according to the present invention;
FIG. 6 is a schematic view of a scene in which a user selects to view a target image in the message processing method of the present invention;
FIG. 7 is a schematic view of a scene where a user clicks an enlarged target area in the message processing method according to the present invention;
fig. 8 is a scene schematic diagram of a preset area and a voice playing interface in a target image interface in the message processing method of the present invention.
The objects, features and advantages of the present invention will be further explained with reference to the accompanying drawings.
Detailed Description
It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention.
In the following description, suffixes such as "module", "component", or "unit" used to denote elements are used only for facilitating the explanation of the present invention, and have no specific meaning in itself. Thus, "module", "component" or "unit" may be used mixedly.
The terminal may be implemented in various forms. For example, the terminal described in the present invention may include a mobile terminal such as a mobile phone, a tablet computer, a notebook computer, a palmtop computer, a Personal Digital Assistant (PDA), a Portable Media Player (PMP), a navigation device, a wearable device, a smart band, a pedometer, and the like, and a fixed terminal such as a Digital TV, a desktop computer, and the like.
The following description will be given by way of example of a mobile terminal, and it will be understood by those skilled in the art that the construction according to the embodiment of the present invention can be applied to a fixed type terminal, in addition to elements particularly used for mobile purposes.
Referring to fig. 1, which is a schematic diagram of a hardware structure of a mobile terminal for implementing various embodiments of the present invention, the mobile terminal 100 may include: RF (Radio Frequency) unit 101, WiFi module 102, audio output unit 103, a/V (audio/video) input unit 104, sensor 105, display unit 106, user input unit 107, interface unit 108, memory 109, processor 110, and power supply 111. Those skilled in the art will appreciate that the mobile terminal architecture shown in fig. 1 is not intended to be limiting of mobile terminals, which may include more or fewer components than those shown, or some components may be combined, or a different arrangement of components.
The following describes each component of the mobile terminal in detail with reference to fig. 1:
the radio frequency unit 101 may be configured to receive and transmit signals during information transmission and reception or during a call, and specifically, receive downlink information of a base station and then process the downlink information to the processor 110; in addition, the uplink data is transmitted to the base station. Typically, radio frequency unit 101 includes, but is not limited to, an antenna, at least one amplifier, a transceiver, a coupler, a low noise amplifier, a duplexer, and the like. In addition, the radio frequency unit 101 can also communicate with a network and other devices through wireless communication. The wireless communication may use any communication standard or protocol, including but not limited to GSM (Global System for Mobile communications), GPRS (General Packet Radio Service), CDMA2000(Code Division Multiple Access 2000), WCDMA (Wideband Code Division Multiple Access), TD-SCDMA (Time Division-Synchronous Code Division Multiple Access), FDD-LTE (Frequency Division duplex Long Term Evolution), and TDD-LTE (Time Division duplex Long Term Evolution).
WiFi belongs to short-distance wireless transmission technology, and the mobile terminal can help a user to receive and send e-mails, browse webpages, access streaming media and the like through the WiFi module 102, and provides wireless broadband internet access for the user. Although fig. 1 shows the WiFi module 102, it is understood that it does not belong to the essential constitution of the mobile terminal, and may be omitted entirely as needed within the scope not changing the essence of the invention.
The audio output unit 103 may convert audio data received by the radio frequency unit 101 or the WiFi module 102 or stored in the memory 109 into an audio signal and output as sound when the mobile terminal 100 is in a call signal reception mode, a call mode, a recording mode, a voice recognition mode, a broadcast reception mode, or the like. Also, the audio output unit 103 may also provide audio output related to a specific function performed by the mobile terminal 100 (e.g., a call signal reception sound, a message reception sound, etc.). The audio output unit 103 may include a speaker, a buzzer, and the like.
The a/V input unit 104 is used to receive audio or video signals. The a/V input Unit 104 may include a Graphics Processing Unit (GPU) 1041 and a microphone 1042, the Graphics Processing Unit 1041 Processing image data of a still image or video obtained by an image capturing device (e.g., a camera) in a video capturing mode or an image capturing mode. The processed image frames may be displayed on the display unit 106. The image frames processed by the graphic processor 1041 may be stored in the memory 109 (or other storage medium) or transmitted via the radio frequency unit 101 or the WiFi module 102. The microphone 1042 may receive sounds (audio data) via the microphone 1042 in a phone call mode, a recording mode, a voice recognition mode, or the like, and may be capable of processing such sounds into audio data. The processed audio (voice) data may be converted into a format output transmittable to a mobile communication base station via the radio frequency unit 101 in case of a phone call mode. The microphone 1042 may implement various types of noise cancellation (or suppression) algorithms to cancel (or suppress) noise or interference generated in the course of receiving and transmitting audio signals.
The mobile terminal 100 also includes at least one sensor 105, such as a light sensor, a motion sensor, and other sensors. Specifically, the light sensor includes an ambient light sensor that can adjust the brightness of the display panel 1061 according to the brightness of ambient light, and a proximity sensor that can turn off the display panel 1061 and/or a backlight when the mobile terminal 100 is moved to the ear. As one of the motion sensors, the accelerometer sensor can detect the magnitude of acceleration in each direction (generally, three axes), can detect the magnitude and direction of gravity when stationary, and can be used for applications of recognizing the posture of a mobile phone (such as horizontal and vertical screen switching, related games, magnetometer posture calibration), vibration recognition related functions (such as pedometer and tapping), and the like; as for other sensors such as a fingerprint sensor, a pressure sensor, an iris sensor, a molecular sensor, a gyroscope, a barometer, a hygrometer, a thermometer, and an infrared sensor, which can be configured on the mobile phone, further description is omitted here.
The display unit 106 is used to display information input by a user or information provided to the user. The Display unit 106 may include a Display panel 1061, and the Display panel 1061 may be configured in the form of a Liquid Crystal Display (LCD), an Organic Light-Emitting Diode (OLED), or the like.
The user input unit 107 may be used to receive input numeric or character information and generate key signal inputs related to user settings and function control of the mobile terminal. Specifically, the user input unit 107 may include a touch panel 1071 and other input devices 1072. The touch panel 1071, also referred to as a touch screen, may collect a touch operation performed by a user on or near the touch panel 1071 (e.g., an operation performed by the user on or near the touch panel 1071 using a finger, a stylus, or any other suitable object or accessory), and drive a corresponding connection device according to a predetermined program. The touch panel 1071 may include two parts of a touch detection device and a touch controller. The touch detection device detects the touch direction of a user, detects a signal brought by touch operation and transmits the signal to the touch controller; the touch controller receives touch information from the touch sensing device, converts the touch information into touch point coordinates, sends the touch point coordinates to the processor 110, and can receive and execute commands sent by the processor 110. In addition, the touch panel 1071 may be implemented in various types, such as a resistive type, a capacitive type, an infrared ray, and a surface acoustic wave. In addition to the touch panel 1071, the user input unit 107 may include other input devices 1072. In particular, other input devices 1072 may include, but are not limited to, one or more of a physical keyboard, function keys (e.g., volume control keys, switch keys, etc.), a trackball, a mouse, a joystick, and the like, and are not limited to these specific examples.
Further, the touch panel 1071 may cover the display panel 1061, and when the touch panel 1071 detects a touch operation thereon or nearby, the touch panel 1071 transmits the touch operation to the processor 110 to determine the type of the touch event, and then the processor 110 provides a corresponding visual output on the display panel 1061 according to the type of the touch event. Although the touch panel 1071 and the display panel 1061 are shown in fig. 1 as two separate components to implement the input and output functions of the mobile terminal, in some embodiments, the touch panel 1071 and the display panel 1061 may be integrated to implement the input and output functions of the mobile terminal, and is not limited herein.
The interface unit 108 serves as an interface through which at least one external device is connected to the mobile terminal 100. For example, the external device may include a wired or wireless headset port, an external power supply (or battery charger) port, a wired or wireless data port, a memory card port, a port for connecting a device having an identification module, an audio input/output (I/O) port, a video I/O port, an earphone port, and the like. The interface unit 108 may be used to receive input (e.g., data information, power, etc.) from external devices and transmit the received input to one or more elements within the mobile terminal 100 or may be used to transmit data between the mobile terminal 100 and external devices.
The memory 109 may be used to store software programs and various data, and the memory 109 may be a computer storage medium, and the memory 109 stores the message alert program of the present invention. The memory 109 may mainly include a storage program area and a storage data area, wherein the storage program area may store an operating system, an application program required by at least one function (such as a sound playing function, an image playing function, etc.), and the like; the storage data area may store data (such as audio data, a phonebook, etc.) created according to the use of the cellular phone, and the like. Further, the memory 109 may include high speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other volatile solid state storage device.
The processor 110 is a control center of the mobile terminal, connects various parts of the entire mobile terminal using various interfaces and lines, and performs various functions of the mobile terminal and processes data by operating or executing software programs and/or modules stored in the memory 109 and calling data stored in the memory 109, thereby performing overall monitoring of the mobile terminal. Such as processor 110, executes a message alert program stored in memory 109 to implement the steps of various embodiments of the message alert method of the present invention.
Processor 110 may include one or more processing units; alternatively, the processor 110 may integrate an application processor, which mainly handles operating systems, user interfaces, application programs, etc., and a modem processor, which mainly handles wireless communications. It will be appreciated that the modem processor described above may not be integrated into the processor 110.
The mobile terminal 100 may further include a power supply 111 (e.g., a battery) for supplying power to various components, and optionally, the power supply 111 may be logically connected to the processor 110 through a power management system, so as to implement functions of managing charging, discharging, and power consumption through the power management system.
Although not shown in fig. 1, the mobile terminal 100 may further include a bluetooth module or the like, which is not described in detail herein.
In order to facilitate understanding of the embodiments of the present invention, a communication network system on which the mobile terminal of the present invention is based is described below.
Referring to fig. 2, fig. 2 is an architecture diagram of a communication Network system according to an embodiment of the present invention, where the communication Network system is an LTE system of a universal mobile telecommunications technology, and the LTE system includes a UE (User Equipment) 201, an E-UTRAN (Evolved UMTS Terrestrial Radio Access Network) 202, an EPC (Evolved Packet Core) 203, and an IP service 204 of an operator, which are in communication connection in sequence.
Specifically, the UE201 may be the terminal 100 described above, and is not described herein again.
The E-UTRAN202 includes eNodeB2021 and other eNodeBs 2022, among others. Among them, the eNodeB2021 may be connected with other eNodeB2022 through backhaul (e.g., X2 interface), the eNodeB2021 is connected to the EPC203, and the eNodeB2021 may provide the UE201 access to the EPC 203.
The EPC203 may include an MME (Mobility Management Entity) 2031, an HSS (Home Subscriber Server) 2032, other MMEs 2033, an SGW (Serving gateway) 2034, a PGW (PDN gateway) 2035, and a PCRF (Policy and Charging Rules Function) 2036, and the like. The MME2031 is a control node that handles signaling between the UE201 and the EPC203, and provides bearer and connection management. HSS2032 is used to provide registers to manage functions such as home location register (not shown) and holds subscriber specific information about service characteristics, data rates, etc. All user data may be sent through SGW2034, PGW2035 may provide IP address assignment for UE201 and other functions, and PCRF2036 is a policy and charging control policy decision point for traffic data flow and IP bearer resources, which selects and provides available policy and charging control decisions for a policy and charging enforcement function (not shown).
The IP services 204 may include the internet, intranets, IMS (IP Multimedia Subsystem), or other IP services, among others.
Although the LTE system is described as an example, it should be understood by those skilled in the art that the present invention is not limited to the LTE system, but may also be applied to other wireless communication systems, such as GSM, CDMA2000, WCDMA, TD-SCDMA, and future new network systems.
Based on the above mobile terminal hardware structure and communication network system, the message processing method of the present invention is proposed in various embodiments.
The invention provides a message processing method, which is mainly applied to mobile terminal equipment and comprises the following steps:
receiving a message sent by a terminal, and acquiring an image and voice in the message;
detecting the type of an operation instruction of a user;
if the operation instruction is an image operation instruction, acquiring a target image in each image, and amplifying the target image to acquire a target image interface, wherein different voice playing interfaces are respectively arranged in each preset area of the target image interface;
and when the preset area is detected to have a target area which is being amplified, playing a first voice corresponding to the target area in the voices through a voice playing interface of the target area.
Referring to fig. 3, the present invention provides a message processing method, in an embodiment of the message processing method, the message processing method includes the following steps:
step S10, receiving the message sent by the terminal, and acquiring the image and voice in the message;
in the embodiment of the invention, the mobile terminal can detect whether each application program receives a message in real time, and when detecting that the application program receives the message, the mobile terminal also needs to determine whether the received message contains images and voice, and when determining that the images and the voice are both sometimes, the mobile terminal can automatically acquire the images and the voice in the message. It should be noted that the application program needs to be an application having a function of receiving a message in the terminal, and may be an application that needs to monitor received messages and is preset by a user, such as WeChat, QQ, or nailing. Determining whether the received message contains the image and the voice means that whether the received message contains the image and the voice is judged within a preset time period set in advance by the user based on the preference requirement of the user.
Step S20, detecting the type of the operation instruction of the user;
when detecting that an application receives images and voices sent by other terminals in the mobile terminal, the type of an operation instruction of a user is required to be detected, namely, whether the user wants to listen to the voices or to view pictures is determined, when the operation instruction of the user is detected to be the image operation instruction, the image corresponding to the image operation instruction is automatically opened and amplified, when the user views the images and amplifies one area in the images, voice information corresponding to the area is automatically played, when the user amplifies the other area, the original voice playing is automatically stopped, and the voice information corresponding to the area is played instead.
However, when the operation instruction of the user is detected to be a voice opening instruction, the voice which the user wants to listen to is automatically opened, and when the voice information triggering the image opening is detected in the voice which is played, the image sent by other terminals is automatically opened.
The starting mode for detecting the operation instruction of the user can be that the user starts in a voice mode, and the mode needs the mobile terminal to have a voice control function. For example, when the user speaks to turn on the first voice, the terminal automatically plays the corresponding voice, and when the user speaks to turn on the second picture, the terminal automatically turns on and enlarges the picture. Or the mobile terminal can be opened in a conventional click-to-touch mode, and the method needs the mobile terminal to have a click-to-touch function. For example, as shown in fig. 6, when the mobile terminal receives two pieces of voice information and one piece of image information, when the user touches and clicks on an image, the user will automatically enter the image interface and enlarge the first image for the user to view.
Step S30, if the operation instruction is an image operation instruction, acquiring a target image in each image, and amplifying the target image to acquire a target image interface, wherein different voice playing interfaces are respectively arranged in each preset area of the target image interface;
the preset area may be a fixed area set in advance by the user. When the operation instruction of the user is found to be an image operation instruction through judgment in the mobile terminal, the target image in all the received images can be determined through the image operation instruction, the target image is amplified, so that a target image interface is obtained, different voice playing interfaces are respectively arranged in each preset area in the target image interface, and when the condition is met, the voice received by the mobile terminal can be played through the voice playing interfaces. It should be noted that the same voice can be played in different voice playing interfaces.
In order to assist understanding of the present invention, different voice playing interfaces are set in each preset area on the target image interface, which is described below by way of example.
For example, as shown in fig. 8, when there are 5 preset areas in the enlarged target image interface, which are an a area, a B area, a C area, a D area, and an F area, respectively, and the played voices corresponding to the voice playing interfaces in the 5 preset areas are different, when there are 3 received voices, which are voice 1, voice 2, and voice 3, respectively, it may be set to play voice 1 in the a area, voice 2 and voice 1 in the B area, voice 3 in the C area, voice 1 and voice 3 in the D area, and voice 1, voice 2, and voice 3 in the F area. It should be noted that the specific setting mode may be set by the user who sends the image and the voice, and is not limited herein.
Step S40, when it is detected that a target area being amplified exists in the preset area, playing a first voice corresponding to the target area in the voices through a first voice playing interface of the target area.
The target area may be an area arbitrarily selected by the user among the respective preset areas based on the preference demand of the user. The first voice playing interface may be a voice playing interface corresponding to the target area, and the user plays the voice related to the target area. The first voice may be a voice that needs to be played in the target area. When a user selects one of the preset areas as a target area to perform amplification processing in the target image interface, the voice playing interface is automatically triggered at the moment due to the fact that the voice playing interface is arranged in the target area in advance, and then a first voice matched with the target area is screened out from the voices through the voice playing interface. For example, when the mobile terminal receives each voice and image sent by other terminals, and after the user clicks the image to enter the image interface, each voice playing interface in the image interface is not triggered, but when the user selects and enlarges a certain area in the image interface, and the area is just provided with the voice playing interface, at this time, the voice playing interface is triggered, and the voice corresponding to the area is automatically screened out from each received voice and played.
In order to assist understanding of the present invention, when the user selects the target image for viewing and plays the corresponding first voice, the following description is given by way of example.
For example, as shown in fig. 7, when the display interface in the mobile terminal is a human portrait interface, and the user wants to view the wrist portion of the person in the image at this time, the user clicks the wrist portion image, at this time, the image of the wrist portion is automatically enlarged and displayed for the user to view, and when the area just has the voice playing interface, the voice corresponding to the wrist portion is also automatically played.
In the embodiment, a message sent by a receiving terminal is used for acquiring images and voice in the message; detecting the type of an operation instruction of a user; if the operation instruction is an image operation instruction, acquiring a target image in each image, and amplifying the target image to acquire a target image interface, wherein different voice playing interfaces are respectively arranged in each preset area of the target image interface; and when the preset area is detected to have a target area which is being amplified, playing a first voice corresponding to the target area in the voices through a first voice playing interface of the target area. When the target area is enlarged in the target image interface, the first voice corresponding to the target area is automatically played, and the user does not need to manually start the corresponding voice, so that the working efficiency is improved for the user, the intelligent effect of the user is also improved, the user can listen to the corresponding voice when checking the image in the received message, the understanding of the user to the image is accelerated, the use experience of the user is improved, the technical effect that the user can automatically play the corresponding voice when checking the image is achieved, and the technical problem that the checking time sequence of the voice and the checking time sequence of the image cannot be flexibly coordinated in the prior art is also solved.
Further, on the basis of the first embodiment of the present invention, a second embodiment of the message processing method of the present invention is provided, where this embodiment is step S40 of the first embodiment of the present invention, and when it is detected that a target area which is being amplified exists in the preset area, a step of playing a first voice corresponding to the target area in the voice through a voice playing interface of the target area is detailed, referring to fig. 4, where the step includes:
step S41, when detecting that there is a target area in the preset area, judging whether the target area has a limiting condition;
in the target image interface, the mobile terminal monitors the operating state of each preset area at any time, and when a certain area (namely a target area) is detected in each preset area to receive an amplification processing command of a user and the target area is amplified, whether the target area has a limitation condition or not can be judged first, and whether corresponding voice is played through a voice playing interface of the target area or not can be determined according to the judgment result. The limiting conditions can be the level of the user account, the number of login days of the account, and the like, are not particularly limited, and can be set by the user based on the preference requirement of the user.
Step S42, if there is no restriction condition, playing a first voice corresponding to the target area in the voices through the voice playing interface of the target area.
When the target area is found to have no limitation condition through judgment, namely anyone can listen to the first voice played through the voice playing interface of the target area when amplifying the target area.
In the embodiment, whether the corresponding voice is played or not is determined by determining whether the target area which is being checked by the user has the limitation condition, so that the safety of the message is improved, and the use experience of the user is also improved.
Specifically, after the step of determining whether the target area has the limitation condition, the method includes:
step S43, if the restriction condition exists, acquiring the account information of the user, and judging whether the account information meets the restriction condition;
when the limited condition exists in the target area through judgment, the account information of the user viewing the target image at the moment needs to be acquired, and then whether the requirement meeting the limited condition exists in the account information is determined. For example, when the limitation condition is that the member can play the voice, it may be determined whether the account information includes the member information, that is, whether the account is a member is determined, and when the account is determined to be a member, it may be determined that the account information satisfies the limitation condition, so that the first voice may be played through the voice playing interface of the target area.
Step S44, if the account information meets the limiting condition, playing a first voice corresponding to the target area in the voice through the voice of the target area;
when the account information meets the limiting condition through judgment, the first voice in each voice can be played through the voice playing interface of the target area, and the order of playing the voices can be that the time of each voice in the first voice received by the mobile terminal is determined first, and then the voices are played in sequence according to the sequence of the time, or the voices are played according to the content of the voices, wherein the specific playing sequence is not limited here.
And step S45, if the account information does not meet the limiting conditions, automatically shielding the voice playing interface of the target area.
When the account information is found to not meet the limiting conditions through judgment, the voice playing interface of the target area can be automatically shielded, namely, when the user enlarges the target area to view the picture, the voice can not be played simultaneously.
In this embodiment, whether to play the corresponding voice for the user is determined by acquiring the account information of the user and determining whether the account information of the user meets the limiting condition, so that the use experience of the user is improved, and meanwhile, the safety of the information is also ensured.
Specifically, after the step of playing the first voice in the voices through the voice playing interface of the target area, the method includes:
step S60, when it is detected that another preset area in the target image interface corresponding to the target area is enlarged, automatically stopping the current first voice playing, and playing a second voice corresponding to the another preset area through the voice playing interface of the another preset area.
When it is detected that another preset area is amplified in the target image interface, that is, the user has viewed the target area in the target image interface and continues to view another preset area, the first voice playing corresponding to the target area is automatically stopped, but the second voice corresponding to the another preset area is obtained again, and the second voice corresponding to the another preset area is played through the voice player of the another preset area. It should be noted that, when preparing to play the second voice in another preset region, the determination condition is similar to that when playing the first voice corresponding to the target region, and specifically, no description is given here. The method for amplifying the other preset area in the target image interface can be that the target area part is firstly reduced, and then the other preset area is amplified; it is also possible to move directly from the target area portion to another preset area without reducing the target area portion.
In this embodiment, the area that the user views in the target image interface is determined to determine which voice playing interface is used to play the corresponding voice, and the voice that is being played before is automatically turned off, so that the use experience of the user is improved, the understanding of the user on the image is accelerated, and the work efficiency of the user is improved.
Specifically, after the step of respectively setting different voice playing interfaces in each preset area of the target image interface, the method includes:
step S70, obtaining the voice type of each voice,
step S80, obtaining the region type of each preset region, and matching each voice type with each region type to obtain the play-supported voice type of each preset region.
The voice type may be voice content information. The region type may be image information of a preset region in the target image. And matching the voice type with the area type can be to judge whether the content matched with the preset area exists in the voice content information, and if so, the voice can be used as the voice type supported by the preset area to play. After each preset area in the target image interface is provided with each different voice playing interface, it is further required to determine which voices can be played by each voice playing interface, at this time, the voice type of each voice received in the mobile terminal needs to be acquired, then the area type of each preset area in the target image needs to be acquired, and then each voice type is matched with each area type, so that the play-supported voice type corresponding to each preset area is obtained. It should be noted that the same voice can be played in multiple preset areas.
Step S90, determining a target area in each of the preset areas, acquiring a first voice type supported by the target area for playing, and using the first voice type as a first voice to be played.
The target area is determined in each preset area of the target image interface, and because the play-supporting voice types corresponding to each preset area are set in the mobile terminal in advance, the first voice type supporting play corresponding to the target area can be determined at this time, and the first voice type can be used as the first voice to be played.
In this embodiment, the third voices corresponding to the preset areas in the target image interface are determined, so that each preset area is guaranteed to have the correct voice corresponding to the preset area, and therefore, the user can play the correct voice when looking over the preset areas, and the user experience is improved.
Further, on the basis of any one of the first to second embodiments of the present invention, a third embodiment of the message processing method of the present invention is proposed, where this embodiment is step S20 of the first embodiment of the present invention, and after the step of detecting the type of the operation instruction of the first user, the method includes:
step A10, if the operation instruction is a voice opening instruction, opening a target voice in each voice based on the voice opening instruction;
the target voice can be a voice corresponding to the opening voice instruction. And when the operation instruction is found to be a voice starting instruction through judgment, the target voice in each voice can be screened out through the operation instruction and played.
Step A20, detecting whether voice information triggering the sub-image in the image exists in the target voice;
the sub-image may be any one of all images that have been received in the mobile terminal. When the target voice is played, whether voice information triggering each image exists in the played content is monitored and detected all the time, and when the voice information triggering one image exists in the played voice information, the image is automatically started and amplified; however, when the voice information of the trigger image is not detected, the target voice continues to be played without any processing on the image.
Step A30, if the voice information triggering the sub-image exists in the target voice, automatically opening and magnifying the sub-image.
When the voice information triggering one image in the images exists in the played target voice detected in the mobile terminal, the primary image is automatically started and amplified. However, when the voice information of the trigger sub-image is not detected in the voice information of the target voice, the target voice continues to be played without triggering the image to be turned on. For example, when the image received by the mobile terminal is image a and the voice information of the triggered image a is image B, when the mobile terminal plays the received voice C, the voice content played by the voice C is constantly monitored and detected, and when the content B in the voice content is detected, the image a is automatically opened and enlarged for the user to view.
In the embodiment, whether the corresponding image is started or not is determined by determining whether the voice information triggering the image exists in the target voice being played, so that the intelligent effect is improved, and the use experience of a user is improved.
Specifically, if the operation instruction is a voice start instruction, after the step of starting the target voice in each voice based on the voice start instruction, the method includes:
step a40, acquiring image information of each image;
step A50, determining the voice information triggering the sub-image in the image according to each image information, and detecting the voice information in the target voice.
The image information may include general content information of the image. When receiving each image sent by other terminals in the mobile terminal, the mobile terminal needs to acquire the image information of the images, and selects the keyword information triggering each image from the image information, namely each image has the keyword information corresponding to the image, and then the keyword information is used as the voice information triggering each image, namely when the target voice is played and certain keyword information in the target voice is detected, the image corresponding to the keyword is automatically acquired, and the image is opened and amplified.
In the embodiment, the voice information for triggering the image opening is determined according to the image information acquired by the mobile terminal, so that the image can be automatically opened without manual opening by a user, and the use experience of the user is improved.
In addition, referring to fig. 5, an embodiment of the present invention further provides a message processing apparatus, where the message processing apparatus includes:
the receiving module is used for receiving a message sent by a terminal and acquiring an image and voice in the message;
the detection module is used for detecting the type of an operation instruction of a user;
the acquisition module is used for acquiring a target image in each image and amplifying the target image to acquire a target image interface if the operation instruction is an image operation instruction, wherein different voice playing interfaces are respectively arranged in each preset area of the target image interface;
and the playing module is used for playing a first voice corresponding to the target area in the voices through a voice playing interface of the target area when detecting that the target area which is being amplified exists in the preset area.
Optionally, the playing module is further configured to:
when detecting that a target area which is being amplified exists in the preset area, judging whether the target area has a limiting condition;
and if the limitation condition does not exist, playing a first voice corresponding to the target area in the voices through a voice playing interface of the target area.
Optionally, the playing module is further configured to:
if the limiting condition exists, acquiring account information of the user, and judging whether the account information meets the limiting condition;
if the account information meets the limiting condition, playing a first voice corresponding to the target area in the voice through a voice playing interface of the target area;
and if the account information does not meet the limiting conditions, automatically shielding the voice playing interface of the target area.
Optionally, the message processing apparatus further includes:
and when the amplification of another preset area in the target image interface relative to the target area is detected, automatically stopping the current first voice playing, and playing a second voice corresponding to the other preset area through a voice playing interface of the other preset area.
Optionally, the message processing apparatus further includes:
the voice type of each of the voices is acquired,
acquiring the region type of each preset region, and matching each voice type with each region type to acquire the play-supporting voice type of each preset region;
and determining a target area in each preset area, acquiring a first voice type supported by the target area to be played, and taking the first voice type as a first voice to be played.
Optionally, the message processing apparatus further includes:
if the operation instruction is a voice opening instruction, opening target voice in the voices based on the voice opening instruction;
detecting whether voice information triggering the sub-image in the image exists in the target voice;
and if the target voice has the voice information triggering the subimage, automatically starting and amplifying the subimage.
Optionally, the message processing apparatus further includes:
acquiring image information of each image;
and determining and triggering voice information of the sub-images in each image according to each image information, and detecting the voice information in the target voice.
The steps implemented by each functional module of the message processing apparatus may refer to each embodiment of the message processing method of the present invention, and are not described herein again.
The present invention also provides a terminal, including: a memory, a processor, a communication bus, and a message handling program stored on the memory:
the communication bus is used for realizing connection communication between the processor and the memory;
the processor is configured to execute the message processing program to implement the steps of the embodiments of the message processing method.
The present invention also provides a storage medium storing one or more programs, which are also executable by one or more processors for implementing the steps of the embodiments of the message processing method described above.
The specific implementation of the storage medium of the present invention is basically the same as the embodiments of the message processing method described above, and is not described herein again.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or system that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or system. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other like elements in a process, method, article, or system that comprises the element.
The above-mentioned serial numbers of the embodiments of the present invention are merely for description and do not represent the merits of the embodiments.
Through the above description of the embodiments, those skilled in the art will clearly understand that the method of the above embodiments can be implemented by software plus a necessary general hardware platform, and certainly can also be implemented by hardware, but in many cases, the former is a better implementation manner. Based on such understanding, the technical solution of the present invention may be embodied in the form of a software product, which is stored in a storage medium (e.g., ROM/RAM, magnetic disk, optical disk) as described above and includes instructions for enabling a terminal device (e.g., a mobile phone, a computer, a server, an air conditioner, or a network device) to execute the method according to the embodiments of the present invention.
The above description is only a preferred embodiment of the present invention, and not intended to limit the scope of the present invention, and all modifications of equivalent structures and equivalent processes, which are made by using the contents of the present specification and the accompanying drawings, or directly or indirectly applied to other related technical fields, are included in the scope of the present invention.

Claims (10)

1. A message processing method, characterized in that the message processing method comprises the steps of:
receiving a message sent by a terminal, and acquiring an image and voice in the message;
detecting the type of an operation instruction of a user;
if the operation instruction is an image operation instruction, acquiring a target image in each image, and amplifying the target image to acquire a target image interface, wherein different voice playing interfaces are respectively arranged in each preset area of the target image interface;
acquiring the voice type supporting playing of each preset area;
determining a target area in each preset area, acquiring a first voice type supported by the target area to be played, and determining a first voice to be played in the voice according to the first voice type;
and when the preset area is detected to have a target area which is being amplified, playing the first voice corresponding to the target area through a voice playing interface of the target area.
2. The message processing method according to claim 1, wherein the step of playing the first voice corresponding to the target area through a voice playing interface of the target area when detecting that the target area being amplified exists in the preset area comprises:
when detecting that a target area which is being amplified exists in the preset area, judging whether the target area has a limiting condition;
and if the limitation condition does not exist, playing the first voice corresponding to the target area through a voice playing interface of the target area.
3. The message processing method of claim 2, wherein the step of determining whether the restriction condition exists in the target area comprises, after:
if the limiting condition exists, acquiring account information of the user, and judging whether the account information meets the limiting condition;
if the account information meets the limiting condition, playing the first voice corresponding to the target area through a voice playing interface of the target area;
and if the account information does not meet the limiting conditions, automatically shielding the voice playing interface of the target area.
4. The message processing method according to claim 1, wherein after the step of playing the first voice corresponding to the target area through the voice playing interface of the target area, the method comprises:
and when the amplification of another preset area in the target image interface relative to the target area is detected, automatically stopping the current first voice playing, and playing a second voice corresponding to the other preset area through a voice playing interface of the other preset area.
5. The message processing method according to claim 1, wherein the step of obtaining the supported playing voice type of each of the preset regions comprises:
the voice type of each of the voices is acquired,
acquiring the region type of each preset region,
and matching each voice type with each area type to obtain the play-supported voice type of each preset area.
6. The message processing method according to claim 1, wherein the step of detecting the type of the operation instruction of the user is followed by:
if the operation instruction is a voice starting instruction, starting a target voice in the voice based on the voice starting instruction;
detecting whether voice information triggering the sub-images in each image exists in the target voice;
and if the target voice has the voice information triggering the subimage, automatically starting and amplifying the subimage.
7. The message processing method according to claim 6, wherein if the operation command is a voice turn-on command, the step of turning on the target voice in the voices based on the voice turn-on command comprises:
acquiring image information of each image;
and determining and triggering voice information of the sub-images in each image according to each image information, and detecting the voice information in the target voice.
8. A message processing apparatus, characterized in that the message processing apparatus comprises:
the receiving module is used for receiving a message sent by a terminal and acquiring an image and voice in the message;
the acquisition module is used for detecting the type of an operation instruction of a user;
the device comprises a setting module, a processing module and a display module, wherein the setting module is used for acquiring a target image in each image and amplifying the target image to acquire a target image interface if the operation instruction is an image operation instruction, and different voice playing interfaces are respectively arranged in each preset area of the target image interface; the voice playing device is also used for acquiring the play-supporting voice types of the preset areas, determining a target area in the preset areas, acquiring a first voice type supported by the target area for playing, and determining a first voice to be played according to the first voice type;
and the playing module is used for playing the first voice corresponding to the target area through a voice playing interface of the target area when detecting that the target area which is subjected to amplification processing exists in the preset area.
9. A mobile terminal, characterized in that the mobile terminal comprises: memory, a processor and a message handling program stored on the memory and executable on the processor, the message handling program when executed by the processor implementing the steps of the message handling method according to any of claims 1 to 7.
10. A storage medium, characterized in that the storage medium has stored thereon a message processing program which, when executed by a processor, implements the steps of the message processing method according to any one of claims 1 to 7.
CN201811462250.1A 2018-11-30 2018-11-30 Message processing method, device, mobile terminal and storage medium Active CN109309762B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811462250.1A CN109309762B (en) 2018-11-30 2018-11-30 Message processing method, device, mobile terminal and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811462250.1A CN109309762B (en) 2018-11-30 2018-11-30 Message processing method, device, mobile terminal and storage medium

Publications (2)

Publication Number Publication Date
CN109309762A CN109309762A (en) 2019-02-05
CN109309762B true CN109309762B (en) 2021-08-10

Family

ID=65223674

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811462250.1A Active CN109309762B (en) 2018-11-30 2018-11-30 Message processing method, device, mobile terminal and storage medium

Country Status (1)

Country Link
CN (1) CN109309762B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111935488B (en) * 2019-05-13 2022-10-28 阿里巴巴集团控股有限公司 Data processing method, information display method, device, server and terminal equipment
CN111147929B (en) * 2019-12-10 2023-04-18 西安万像电子科技有限公司 Application mode switching method and device and terminal
CN111128139B (en) * 2019-12-18 2022-07-08 思必驰科技股份有限公司 Non-invasive voice test method and device

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9325776B2 (en) * 2013-01-08 2016-04-26 Tangome, Inc. Mixed media communication
CN107895006B (en) * 2017-11-07 2021-03-02 Oppo广东移动通信有限公司 Audio playing method and device, storage medium and electronic equipment

Also Published As

Publication number Publication date
CN109309762A (en) 2019-02-05

Similar Documents

Publication Publication Date Title
CN109151169B (en) Camera authority management method, mobile terminal and computer readable storage medium
CN109165074B (en) Game screenshot sharing method, mobile terminal and computer-readable storage medium
CN112799577B (en) Method, terminal and storage medium for projecting small window
CN107807767B (en) Communication service processing method, terminal and computer readable storage medium
CN109068167B (en) Screen recording interaction method, mobile terminal, wearable device and storage medium
CN107635070B (en) Message reminding method, terminal and storage medium
CN110187808B (en) Dynamic wallpaper setting method and device and computer-readable storage medium
CN109195213B (en) Mobile terminal screen control method, mobile terminal and computer readable storage medium
CN107862217B (en) Position information acquisition method, mobile terminal and computer storage medium
CN107832032B (en) Screen locking display method and mobile terminal
CN109375846B (en) Method and device for displaying breathing icon, mobile terminal and readable storage medium
CN109309762B (en) Message processing method, device, mobile terminal and storage medium
CN108958936B (en) Application program switching method, mobile terminal and computer readable storage medium
CN107682541B (en) Audio control method for screen projection, mobile terminal and storage medium
CN108156325B (en) Double-sided screen incoming call control method and device and computer readable storage medium
CN107422956B (en) Mobile terminal operation response method, mobile terminal and readable storage medium
CN108282608B (en) Multi-region focusing method, mobile terminal and computer readable storage medium
CN107239208B (en) Method, apparatus, and computer-readable storage medium for processing screenshot
CN113326012A (en) Processing method, mobile terminal and storage medium
CN112604281A (en) Game visual field control method, mobile terminal and computer readable storage medium
CN110083294B (en) Screen capturing method, terminal and computer readable storage medium
CN109683796B (en) Interaction control method, equipment and computer readable storage medium
CN108418966B (en) Message prompting method, mobile terminal and computer readable storage medium
CN108762709B (en) Terminal control method, terminal and computer readable storage medium
CN107678622B (en) Application icon display method, terminal and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant