CN116168701A - Interaction method, intelligent terminal and storage medium - Google Patents
Interaction method, intelligent terminal and storage medium Download PDFInfo
- Publication number
- CN116168701A CN116168701A CN202310168271.7A CN202310168271A CN116168701A CN 116168701 A CN116168701 A CN 116168701A CN 202310168271 A CN202310168271 A CN 202310168271A CN 116168701 A CN116168701 A CN 116168701A
- Authority
- CN
- China
- Prior art keywords
- target application
- virtual object
- application
- interaction
- dynamic effect
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 230000003993 interaction Effects 0.000 title claims abstract description 65
- 238000000034 method Methods 0.000 title claims abstract description 56
- 230000000694 effects Effects 0.000 claims abstract description 47
- 230000002452 interceptive effect Effects 0.000 claims abstract description 17
- 238000004590 computer program Methods 0.000 claims description 14
- 230000004044 response Effects 0.000 claims description 14
- 230000008859 change Effects 0.000 claims description 12
- 230000033001 locomotion Effects 0.000 claims description 9
- 230000004660 morphological change Effects 0.000 claims description 9
- 238000010612 desalination reaction Methods 0.000 claims description 2
- 230000006870 function Effects 0.000 description 20
- 230000006854 communication Effects 0.000 description 12
- 238000004891 communication Methods 0.000 description 11
- 230000008569 process Effects 0.000 description 11
- 230000006399 behavior Effects 0.000 description 7
- 238000001514 detection method Methods 0.000 description 5
- 238000010586 diagram Methods 0.000 description 5
- 238000007726 management method Methods 0.000 description 5
- 230000001960 triggered effect Effects 0.000 description 5
- 238000010295 mobile communication Methods 0.000 description 4
- 238000012545 processing Methods 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 3
- 230000003542 behavioural effect Effects 0.000 description 2
- 239000002131 composite material Substances 0.000 description 2
- 239000004973 liquid crystal related substance Substances 0.000 description 2
- 230000007774 longterm Effects 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 230000005236 sound signal Effects 0.000 description 2
- 230000001133 acceleration Effects 0.000 description 1
- 238000013473 artificial intelligence Methods 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000007599 discharging Methods 0.000 description 1
- 238000005562 fading Methods 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 230000005484 gravity Effects 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000011664 signaling Effects 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 230000001629 suppression Effects 0.000 description 1
- 238000010897 surface acoustic wave method Methods 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/223—Execution procedure of a spoken command
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/225—Feedback of the input speech
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D10/00—Energy efficient computing, e.g. low power processors, power management or thermal management
Landscapes
- Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
The application provides an interaction method, an intelligent terminal and a storage medium, wherein the method comprises the following steps: s11: responding to the awakening operation and outputting a virtual object; s12: and responding to the interactive operation, and confirming or generating the dynamic effect of the target application based on the virtual object. The interaction scheme provided by the application can enrich the interaction modes of the intelligent voice assistant and the user.
Description
Technical Field
The application relates to the technical field of interaction, in particular to an interaction method, an intelligent terminal and a storage medium.
Background
Along with the development of artificial intelligence technology, various intelligent voice assistant products appear in the market, and the intelligent voice assistant has the function of carrying out voice interaction with the user, can strengthen the user and to apply the goodness of product and strengthen user's viscidity, can promote user experience to a great extent.
In the process of designing and implementing the present application, the inventors found that at least the following problems exist: the current intelligent voice assistant has a single interaction mode with the user.
The foregoing description is provided for general background information and does not necessarily constitute prior art.
Disclosure of Invention
Aiming at the technical problems, the application provides an interaction method, an intelligent terminal and a storage medium, which can enrich interaction modes of an intelligent voice assistant and a user.
The application provides an interaction method, which comprises the following steps:
s11: responding to the awakening operation and outputting a virtual object;
s12: and responding to the interactive operation, and confirming or generating the dynamic effect of the target application based on the virtual object.
Optionally, the step S12 includes:
identifying or determining a target application corresponding to the interactive operation;
and confirming or generating dynamic effects of the target application based on the virtual object according to the position information of the target application.
Optionally, the determining or generating the dynamic effect of the target application based on the virtual object according to the location information of the target application includes:
establishing a moving path of the virtual object and the target application according to the position information of the target application;
after the virtual object is controlled to move to the target application according to the moving path, starting the target application;
and triggering the virtual object to perform morphological change when the target application is started.
Optionally, when the target application is started, triggering the virtual object to perform morphological change includes:
Acquiring a first form and a second form of the virtual object, wherein the first form corresponds to an initial state of the target application, and the second form corresponds to a starting state of the target application;
and triggering the virtual object to change from the first form to the second form when the target application is switched from the initial state to the starting state.
Optionally, the method further comprises:
and triggering the virtual object to change from the second form to the first form in response to an application exit operation aiming at the target application.
Optionally, the determining or generating the dynamic effect of the target application based on the virtual object according to the location information of the target application includes:
acquiring a preset interaction point;
and respectively controlling the target application and the virtual object to move towards the interaction point, and displaying the moving effect after the movement.
Optionally, the controlling the target application and the virtual object to move towards the interaction point respectively and displaying the moving effect after the movement includes:
identifying or determining that the target application and virtual object are located at the interaction point;
amplifying the application icon of the target application at a preset position of the virtual object;
and identifying or determining that the application icon is enlarged to a preset proportion, and displaying an application interface of the target application.
Optionally, after the step S12, the method further includes:
and responding to the hiding operation aiming at the target application, and outputting the dynamic effect of hiding the target application by the virtual object.
Optionally, the responding to the hiding operation for the target application outputs the dynamic effect of hiding the target application by the virtual object, including at least one of the following:
responding to a minimizing operation aiming at a target application, and outputting dynamic effects of dragging and/or hiding the target application by the virtual object;
and responding to closing operation aiming at the target application, and outputting the virtual object desalination and/or the dynamic effect of closing the target application.
The application also provides an intelligent terminal, including: the system comprises a memory and a processor, wherein the memory stores a computer program which realizes the steps of any interaction method when being executed by the processor.
The present application also provides a storage medium storing a computer program which, when executed by a processor, implements the steps of the interaction method as described in any of the above.
As described above, the interaction method of the present application includes: and responding to the wake-up operation, outputting a virtual object, responding to the interactive operation, and confirming or generating the dynamic effect of the target application based on the virtual object. According to the interaction method, the interaction operation is responded, the dynamic effect of the virtual object for starting the target application is output, and the interaction mode of the intelligent voice assistant and the user can be enriched.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the application and together with the description, serve to explain the principles of the application. In order to more clearly illustrate the technical solutions of the embodiments of the present application, the drawings that are needed in the description of the embodiments will be briefly described below, and it will be obvious to those skilled in the art that other drawings can be obtained from these drawings without inventive effort.
Fig. 1 is a schematic hardware structure of a mobile terminal implementing various embodiments of the present application;
fig. 2 is a schematic diagram of a communication network system according to an embodiment of the present application;
fig. 3 is a schematic view of a scenario of an interaction method provided in an embodiment of the present application;
FIG. 4 is a schematic flow chart of an interaction method provided in an embodiment of the present application;
fig. 5 to 7 are schematic interface diagrams of an interaction method according to an embodiment of the present application;
fig. 8 is a schematic structural diagram of a first implementation of the interaction device provided in the embodiment of the present application.
The realization, functional characteristics and advantages of the present application will be further described with reference to the embodiments, referring to the attached drawings. Specific embodiments thereof have been shown by way of example in the drawings and will herein be described in more detail. These drawings and the written description are not intended to limit the scope of the inventive concepts in any way, but to illustrate the concepts of the present application to those skilled in the art by reference to specific embodiments.
Detailed Description
Reference will now be made in detail to exemplary embodiments, examples of which are illustrated in the accompanying drawings. When the following description refers to the accompanying drawings, the same numbers in different drawings refer to the same or similar elements, unless otherwise indicated. The implementations described in the following exemplary examples are not representative of all implementations consistent with the present application. Rather, they are merely examples of apparatus and methods consistent with some aspects of the present application as detailed in the accompanying claims.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, the element defined by the phrase "comprising one … …" does not exclude the presence of other identical elements in a process, method, article, or apparatus that comprises the element, and furthermore, elements having the same name in different embodiments of the present application may have the same meaning or may have different meanings, a particular meaning of which is to be determined by its interpretation in this particular embodiment or by further combining the context of this particular embodiment.
It should be understood that although the terms first, second, third, etc. may be used herein to describe various information, these information should not be limited by these terms. These terms are only used to distinguish one type of information from another. For example, first information may also be referred to as second information, and similarly, second information may also be referred to as first information, without departing from the scope herein. The word "if" as used herein may be interpreted as "at … …" or "at … …" or "responsive to a determination", depending on the context. Furthermore, as used herein, the singular forms "a", "an" and "the" are intended to include the plural forms as well, unless the context indicates otherwise. It will be further understood that the terms "comprises," "comprising," "includes," and/or "including" specify the presence of stated features, steps, operations, elements, components, items, categories, and/or groups, but do not preclude the presence, presence or addition of one or more other features, steps, operations, elements, components, items, categories, and/or groups. The terms "or," "and/or," "including at least one of," and the like, as used herein, may be construed as inclusive, or meaning any one or any combination. For example, "including at least one of: A. b, C "means" any one of the following: a, A is as follows; b, a step of preparing a composite material; c, performing operation; a and B; a and C; b and C; a and B and C ", again as examples," A, B or C "or" A, B and/or C "means" any of the following: a, A is as follows; b, a step of preparing a composite material; c, performing operation; a and B; a and C; b and C; a and B and C). An exception to this definition will occur only when a combination of elements, functions, steps or operations are in some way inherently mutually exclusive.
It should be understood that, although the steps in the flowcharts in the embodiments of the present application are shown in order as indicated by the arrows, these steps are not necessarily performed in order as indicated by the arrows. The steps are not strictly limited in order and may be performed in other orders, unless explicitly stated herein. Moreover, at least some of the steps in the figures may include multiple sub-steps or stages that are not necessarily performed at the same time, but may be performed at different times, the order of their execution not necessarily occurring in sequence, but may be performed alternately or alternately with other steps or at least a portion of the other steps or stages.
The words "if", as used herein, may be interpreted as "at … …" or "at … …" or "in response to a determination" or "in response to a detection", depending on the context. Similarly, the phrase "if determined" or "if detected (stated condition or event)" may be interpreted as "when determined" or "in response to determination" or "when detected (stated condition or event)" or "in response to detection (stated condition or event), depending on the context.
It should be noted that, in this document, step numbers such as S11 and S12 are adopted, and the purpose of the present invention is to more clearly and briefly describe the corresponding content, and not to constitute a substantial limitation on the sequence, and those skilled in the art may execute S12 first and then execute S11 when implementing the present invention, which is within the scope of protection of the present application.
It should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the present application.
In the following description, suffixes such as "module", "component", or "unit" for representing elements are used only for facilitating the description of the present application, and are not of specific significance per se. Thus, "module," "component," or "unit" may be used in combination.
The intelligent terminal may be implemented in various forms. For example, the smart terminals described in the present application may include mobile terminals such as cell phones, tablet computers, notebook computers, palm computers, personal digital assistants (Personal Digital Assistant, PDA), portable media players (Portable Media Player, PMP), navigation devices, wearable devices, smart bracelets, pedometers, and stationary terminals such as digital TVs, desktop computers, and the like.
The following description will be given taking a mobile terminal as an example, and those skilled in the art will understand that the configuration according to the embodiment of the present application can be applied to a fixed type terminal in addition to elements particularly used for a moving purpose.
Referring to fig. 1, which is a schematic hardware structure of a mobile terminal implementing various embodiments of the present application, the mobile terminal 100 may include: an RF (Radio Frequency) unit 101, a WiFi module 102, an audio output unit 103, an a/V (audio/video) input unit 104, a sensor 105, a display unit 106, a user input unit 107, an interface unit 108, a memory 109, a processor 110, and a power supply 111. Those skilled in the art will appreciate that the mobile terminal structure shown in fig. 1 is not limiting of the mobile terminal and that the mobile terminal may include more or fewer components than shown, or may combine certain components, or a different arrangement of components.
The following describes the components of the mobile terminal in detail with reference to fig. 1:
the radio frequency unit 101 may be used for receiving and transmitting signals during the information receiving or communication process, specifically, after receiving downlink information of the base station, processing the downlink information by the processor 110; and, the uplink data is transmitted to the base station. Typically, the radio frequency unit 101 includes, but is not limited to, an antenna, at least one amplifier, a transceiver, a coupler, a low noise amplifier, a duplexer, and the like. In addition, the radio frequency unit 101 may also communicate with networks and other devices via wireless communications. The wireless communication may use any communication standard or protocol including, but not limited to, GSM (Global System of Mobile communication, global system for mobile communications), GPRS (General Packet Radio Service ), CDMA2000 (Code Division Multiple Access, CDMA 2000), WCDMA (Wideband Code Division Multiple Access ), TD-SCDMA (Time Division-Synchronous Code Division Multiple Access, time Division synchronous code Division multiple access), FDD-LTE (Frequency Division Duplexing-Long Term Evolution, frequency Division duplex long term evolution) and TDD-LTE (Time Division Duplexing-Long Term Evolution, time Division duplex long term evolution), 5G, etc.
WiFi belongs to a short-distance wireless transmission technology, and a mobile terminal can help a user to send and receive e-mails, browse web pages, access streaming media and the like through the WiFi module 102, so that wireless broadband Internet access is provided for the user. Although fig. 1 shows a WiFi module 102, it is understood that it does not belong to the necessary constitution of the mobile terminal, and can be omitted entirely as required within a range that does not change the essence of the application.
The audio output unit 103 may convert audio data received by the radio frequency unit 101 or the WiFi module 102 or stored in the memory 109 into an audio signal and output as sound when the mobile terminal 100 is in a call signal reception mode, a talk mode, a recording mode, a voice recognition mode, a broadcast reception mode, or the like. Also, the audio output unit 103 may also provide audio output (e.g., a call signal reception sound, a message reception sound, etc.) related to a specific function performed by the mobile terminal 100. The audio output unit 103 may include a speaker, a buzzer, and the like.
The a/V input unit 104 is used to receive an audio or video signal. The a/V input unit 104 may include a graphics processor (Graphics Processing Unit, GPU) 1041 and a microphone 1042, the graphics processor 1041 processing image data of still pictures or video obtained by an image capturing device (e.g., a camera) in a video capturing mode or an image capturing mode. The processed image frames may be displayed on the display unit 106. The image frames processed by the graphics processor 1041 may be stored in the memory 109 (or other storage medium) or transmitted via the radio frequency unit 101 or the WiFi module 102. The microphone 1042 can receive sound (audio data) via the microphone 1042 in a phone call mode, a recording mode, a voice recognition mode, and the like, and can process such sound into audio data. The processed audio (voice) data may be converted into a format output that can be transmitted to the mobile communication base station via the radio frequency unit 101 in the case of a telephone call mode. The microphone 1042 may implement various types of noise cancellation (or suppression) algorithms to cancel (or suppress) noise or interference generated in the course of receiving and transmitting the audio signal.
The mobile terminal 100 also includes at least one sensor 105, such as a light sensor, a motion sensor, and other sensors. Optionally, the light sensor includes an ambient light sensor and a proximity sensor, optionally, the ambient light sensor may adjust the brightness of the display panel 1061 according to the brightness of ambient light, and the proximity sensor may turn off the display panel 1061 and/or the backlight when the mobile terminal 100 moves to the ear. As one of the motion sensors, the accelerometer sensor can detect the acceleration in all directions (generally three axes), and can detect the gravity and direction when stationary, and can be used for applications of recognizing the gesture of a mobile phone (such as horizontal and vertical screen switching, related games, magnetometer gesture calibration), vibration recognition related functions (such as pedometer and knocking), and the like; as for other sensors such as fingerprint sensors, pressure sensors, iris sensors, molecular sensors, gyroscopes, barometers, hygrometers, thermometers, infrared sensors, etc. that may also be configured in the mobile phone, the detailed description thereof will be omitted.
The display unit 106 is used to display information input by a user or information provided to the user. The display unit 106 may include a display panel 1061, and the display panel 1061 may be configured in the form of a liquid crystal display (Liquid Crystal Display, LCD), an Organic Light-Emitting Diode (OLED), or the like.
The user input unit 107 may be used to receive input numeric or character information and to generate key signal inputs related to user settings and function control of the mobile terminal. Alternatively, the user input unit 107 may include a touch panel 1071 and other input devices 1072. The touch panel 1071, also referred to as a touch screen, may collect touch operations thereon or thereabout by a user (e.g., operations of the user on the touch panel 1071 or thereabout by using any suitable object or accessory such as a finger, a stylus, etc.) and drive the corresponding connection device according to a predetermined program. The touch panel 1071 may include two parts of a touch detection device and a touch controller. Optionally, the touch detection device detects the touch azimuth of the user, detects a signal brought by touch operation, and transmits the signal to the touch controller; the touch controller receives touch information from the touch detection device, converts it into touch point coordinates, and sends the touch point coordinates to the processor 110, and can receive and execute commands sent from the processor 110. Further, the touch panel 1071 may be implemented in various types such as resistive, capacitive, infrared, and surface acoustic wave. The user input unit 107 may include other input devices 1072 in addition to the touch panel 1071. Alternatively, other input devices 1072 may include, but are not limited to, one or more of a physical keyboard, function keys (e.g., volume control keys, switch keys, etc.), a trackball, mouse, joystick, etc., as specifically not limited herein.
Alternatively, the touch panel 1071 may overlay the display panel 1061, and when the touch panel 1071 detects a touch operation thereon or thereabout, the touch panel 1071 is transferred to the processor 110 to determine the type of touch event, and the processor 110 then provides a corresponding visual output on the display panel 1061 according to the type of touch event. Although in fig. 1, the touch panel 1071 and the display panel 1061 are two independent components for implementing the input and output functions of the mobile terminal, in some embodiments, the touch panel 1071 may be integrated with the display panel 1061 to implement the input and output functions of the mobile terminal, which is not limited herein.
The interface unit 108 serves as an interface through which at least one external device can be connected with the mobile terminal 100. For example, the external devices may include a wired or wireless headset port, an external power (or battery charger) port, a wired or wireless data port, a memory card port, a port for connecting a device having an identification module, an audio input/output (I/O) port, a video I/O port, an earphone port, and the like. The interface unit 108 may be used to receive input (e.g., data information, power, etc.) from an external device and transmit the received input to one or more elements within the mobile terminal 100 or may be used to transmit data between the mobile terminal 100 and an external device.
The processor 110 is a control center of the mobile terminal, connects various parts of the entire mobile terminal using various interfaces and lines, and performs various functions of the mobile terminal and processes data by running or executing software programs and/or modules stored in the memory 109 and calling data stored in the memory 109, thereby performing overall monitoring of the mobile terminal. Processor 110 may include one or more processing units; preferably, the processor 110 may integrate an application processor and a modem processor, the application processor optionally handling mainly an operating system, a user interface, an application program, etc., the modem processor handling mainly wireless communication. It will be appreciated that the modem processor described above may not be integrated into the processor 110.
The mobile terminal 100 may further include a power source 111 (e.g., a battery) for supplying power to the respective components, and preferably, the power source 111 may be logically connected to the processor 110 through a power management system, so as to perform functions of managing charging, discharging, and power consumption management through the power management system.
Although not shown in fig. 1, the mobile terminal 100 may further include a bluetooth module or the like, which is not described herein.
In order to facilitate understanding of the embodiments of the present application, a communication network system on which the mobile terminal of the present application is based will be described below.
Referring to fig. 2, fig. 2 is a schematic diagram of a communication network system provided in the embodiment of the present application, where the communication network system is an LTE system of a general mobile communication technology, and the LTE system includes a UE (User Equipment) 201, an e-UTRAN (Evolved UMTS Terrestrial Radio Access Network ) 202, an epc (Evolved Packet Core, evolved packet core) 203, and an IP service 204 of an operator that are sequentially connected in communication.
Alternatively, the UE201 may be the terminal 100 described above, which is not described here again.
The E-UTRAN202 includes eNodeB2021 and other eNodeB2022, etc. Alternatively, the eNodeB2021 may connect with other enodebs 2022 over a backhaul (e.g., X2 interface), the eNodeB2021 is connected to the EPC203, and the eNodeB2021 may provide access for the UE201 to the EPC 203.
EPC203 may include MME (Mobility Management Entity ) 2031, hss (Home Subscriber Server, home subscriber server) 2032, other MMEs 2033, SGW (Serving Gate Way) 2034, pgw (PDN Gate Way) 2035 and PCRF (Policy and Charging Rules Function, policy and tariff function entity) 2036, and so on. Optionally, MME2031 is a control node that handles signaling between UE201 and EPC203, providing bearer and connection management. HSS1232 is used to provide registers to manage functions such as home location registers (not shown) and to hold user-specific information about service characteristics, data rates, etc. All user data may be sent through SGW2034 and PGW2035 may provide IP address allocation and other functions for UE201, PCRF2036 is a policy and charging control policy decision point for traffic data flows and IP bearer resources, which selects and provides available policy and charging control decisions for a policy and charging enforcement function (not shown).
Although the LTE system is described above as an example, it should be understood by those skilled in the art that the present application is not limited to LTE systems, but may be applied to other wireless communication systems, such as GSM, CDMA2000, WCDMA, TD-SCDMA, 5G, and future new network systems (e.g., 6G), etc.
Based on the above-mentioned mobile terminal hardware structure and communication network system, various embodiments of the present application are presented.
In some implementations, the intelligent terminal installs multiple applications, and the user may launch the applications through the intelligent voice assistant, such as launching application a through the intelligent voice assistant, while using the applications.
In some current schemes, a user wakes up an intelligent voice assistant and speaks an application name of an application a so that the intelligent voice assistant opens the application a, and in the process of opening the application a, the user cannot perceive whether the application a is correctly opened, only when an application interface is displayed, the user can know which application is opened, and the man-machine interaction mode of the whole opening process is single.
In order to solve the above-mentioned existing technical problems, the present application provides an interaction method, optionally, please refer to fig. 3, when a user uses an intelligent terminal, wake up an intelligent voice assistant in the intelligent terminal through a wake-up word, the intelligent terminal responds to the wake-up operation, and outputs a virtual object (i.e. the intelligent voice assistant), the intelligent terminal responds to the user's interaction operation, and confirms or generates the dynamic effect of a target application based on the virtual object.
Referring to fig. 4, fig. 4 is a flow chart of an interaction method according to an embodiment of the present application. The interaction method specifically comprises the following steps:
s11, responding to the awakening operation, and outputting the virtual object.
The wake-up is to wake up a voice assistant in a device sleep state, and generally, a fixed wake-up word is set for voice wake-up (KWS), and after a user speaks the wake-up word, a voice recognition function on a terminal is in a working state, otherwise, the voice assistant is in a sleep state, for example, an acoustic model constructed based on a deep neural network outputs a recognition result, so that voice wake-up is performed according to the recognition result.
Optionally, the virtual object may be a virtual character or a virtual animal, or may be a virtual expression, which may be specifically set according to an actual situation, which is not described herein.
Optionally, in response to a wake-up operation of the user, determining a wake-up word corresponding to the wake-up operation, and then identifying the wake-up word as a preset wake-up word, and displaying the virtual object on the current interface.
And S12, responding to the interactive operation, and confirming or generating the dynamic effect of the target application based on the virtual object.
After the virtual object is output, the user can interact with the intelligent terminal to start the target application, and when the target application is started, the dynamic effect of the target application is confirmed or generated based on the virtual object.
Alternatively, the interaction may be at least one of: air-spaced gestures, voice control, behavioral control, vibration control.
Optionally, the blank gesture finger is operated according to a preset gesture within a range of a certain distance from a screen page of the mobile phone through a finger or a touch pen. Optionally, the blank gesture may be at least one of: space circle, space arc, space semicircle, space straight line, space curve, space check number, space check character, etc.
Optionally, the voice control refers to the user issuing voice control instructions.
Optionally, the behavior control means that when the current behavior of the user accords with the preset behavior, a first operation aiming at the target window can be triggered, and the target information is determined or generated according to the type of the first operation. Alternatively, the behavior may be a behavior in which the user picks up the smart terminal, or the behavior may be a behavior in which the user places the smart terminal on a desktop, or the like.
Optionally, vibration control refers to triggering an opening operation for a target application when the number of vibration times of the intelligent terminal accords with a preset number of vibration times, and/or the vibration frequency of the intelligent terminal accords with a preset vibration frequency, and/or the vibration duration of the intelligent terminal accords with a preset duration, and/or the vibration melody accords with a preset melody, where the target application may be preset by a user.
Alternatively, the target application corresponding to the interaction may be identified, and then, based on the location information of the target application, the corresponding dynamic effect is output, that is, step S12 may specifically include:
identifying or determining a target application corresponding to the interactive operation;
and confirming or generating the dynamic effect of the target application based on the virtual object according to the position information of the target application.
Optionally, the location information of the target application carries the location of the target application on the current interface and the vertex coordinates of the application icon of the target application, and based on the location information, the form of the virtual object may be changed, specifically may be the following manner:
mode one: establishing a moving path based on the position information, and triggering the virtual object to perform morphological change based on the moving path, that is, optionally, in some embodiments, the step of "confirming or generating a dynamic effect of the target application based on the virtual object according to the position information of the target application" specifically may include:
establishing a moving path of the virtual object and the target application according to the position information of the target application;
after the virtual object is controlled to move to the target application according to the moving path, starting the target application;
and triggering the virtual object to perform morphological change when the target application is started.
Optionally, determining a relative position relationship between the virtual object and the target application according to the position information of the target application, and then establishing a moving path of the virtual object and the target application according to the relative position relationship, wherein the moving path can be a straight path or a curved path, specifically can be set according to actual conditions, and then starting the target application after controlling the virtual object to move to the target application according to the moving path, and triggering the virtual object to perform morphological change when the target application is started, wherein the morphological change of the virtual object can be a style change or an object change (such as replacing the virtual object).
Optionally, referring to fig. 5, a relative position relationship between the virtual object D and the target application T is determined according to position information of the target application, then, a linear moving path between the virtual object D and the target application T is established according to the relative position relationship, then, after the virtual object D is controlled to move to the target application T according to the moving path, the target application T is started, and when the target application T is started, the virtual object D is triggered to change from the first form to the second form, such as from the standing posture to the throwing posture in fig. 5.
Optionally, in some embodiments, the step of "triggering the virtual object to perform a morphological change when the target application is opened" may specifically include:
acquiring a first form and a second form of a virtual object;
when the target application is switched from the initial state to the starting state, the virtual object is triggered to change from the first form to the second form.
Optionally, in some embodiments, the first configuration corresponds to an initial state of the target application, the second configuration corresponds to a start state of the target application, and both the first configuration and the second configuration may be preset, for example, the virtual object is a virtual task, the first configuration may be set to a standing posture, and the second configuration may be set to a throwing posture, that is, when the target application is started, the virtual object is switched from the standing posture to the throwing posture to be displayed.
Optionally, in some embodiments, the interaction method of the present application may specifically further include: and triggering the virtual object to change from the second form to the first form in response to the application exit operation aiming at the target application.
Optionally, referring to fig. 6, after the user exits the target application T, the virtual object D is triggered to change from the second configuration to the first configuration.
Mode two: the target application and the virtual object may be controlled to move towards a preset interaction point, and corresponding dynamic effects are displayed, that is, optionally, in some embodiments, the step of "determining or generating, based on the virtual object, the dynamic effects of the target application according to the location information of the target application" may specifically include:
acquiring a preset interaction point;
and respectively controlling the target application and the virtual object to move towards the interaction point, and displaying the moving effect after the movement.
Optionally, the user may preset an interaction point, where the interaction point may be set at any position of the current interface, when the target application is started, the target application and the virtual object are controlled to move toward the interaction point, when the target application and the virtual object are connected at the interaction point, a corresponding dynamic effect is output, and then, the application interface of the target application is switched and displayed.
Optionally, in some embodiments, when the target application and the virtual object are connected at the interaction point, that is, the target application and the virtual object are both located at the interaction point, the application icon of the target application is enlarged, and then the application interface displaying the target application is switched, that is, the step of "controlling the target application and the virtual object to move towards the interaction point and displaying the moving effect after the movement respectively" may specifically include:
Identifying or determining that the target application and the virtual object are located at the interaction point;
amplifying an application icon of the target application at a preset position of the virtual object;
and identifying or determining that the application icons are enlarged to a preset proportion, and displaying an application interface of the target application.
Optionally, referring to fig. 7, it is identified or determined that the target application T and the virtual object D are located at the interaction point, then, an application icon of the target application T is enlarged at a hand of the virtual object D, and when the application icon is enlarged to a preset ratio, an application interface of the target application is displayed.
After the target application is started, the target application may be further hidden, and in order to enrich the interaction mode of the intelligent voice assistant and the user, optionally, in some embodiments, a corresponding dynamic effect may be further output, that is, after step S12, the method may specifically further include: and responding to the hiding operation aiming at the target application, and outputting the dynamic effect of the virtual object hiding the target application.
Optionally, the hiding operation may include a minimizing operation and/or a closing operation, and the minimizing operation for the target application may hide an application interface of the target application, but does not affect the operation of the target application; and for the closing operation of the target application, the process of the target application may be closed.
Optionally, in response to the minimizing operation for the target application T, the dynamic effect of the virtual object D dragging and/or hiding the target application is output, and in response to the closing operation for the target application T, the dynamic effect of the virtual object D fading and/or closing the target application D is output.
According to the interactive method, the virtual object is output in response to the awakening operation, the interactive operation is responded, the dynamic effect of the target application is confirmed or generated based on the virtual object, and through the technical scheme, when the target application is started, the interactive operation can be responded, the dynamic effect of the virtual object for starting the target application is output, and the interactive mode of the intelligent voice assistant and the user can be enriched.
Referring to fig. 8, fig. 8 is a schematic structural diagram of the interaction device provided in the present application, where the interaction device 30 may be integrated in or be an intelligent terminal, and may specifically include a first output module 301 and a second output module 302, which are specifically as follows:
the first output module 301 may be configured to output the virtual object in response to a wake operation.
Optionally, the virtual object may be a virtual character or a virtual animal, or may be a virtual expression, which may be specifically set according to an actual situation, which is not described herein.
Optionally, in response to a wake-up operation of the user, determining a wake-up word corresponding to the wake-up operation, and then identifying the wake-up word as a preset wake-up word, and displaying the virtual object on the current interface.
The second output module 302 may be configured to confirm or generate a dynamic effect of the target application based on the virtual object in response to the interaction.
After the virtual object is output, the user can interact with the intelligent terminal to start the target application, and when the target application is started, the dynamic effect of the target application is confirmed or generated based on the virtual object.
Alternatively, the interaction may be at least one of: air-spaced gestures, voice control, behavioral control, vibration control.
Optionally, the second output module 302 may identify the target application corresponding to the interaction, and then, based on the location information of the target application, output a corresponding dynamic effect, that is, the second output module 302 may specifically be configured to: identifying or determining a target application corresponding to the interactive operation; and confirming or generating the dynamic effect of the target application based on the virtual object according to the position information of the target application.
Alternatively, the second output module 302 may be specifically configured to: establishing a moving path of the virtual object and the target application according to the position information of the target application; and after the virtual object is controlled to move to the target application according to the moving path, triggering the virtual object to change the form when the target application is started.
Alternatively, the second output module 302 may be specifically configured to: acquiring a first form and a second form of a virtual object; when the target application is switched from the initial state to the starting state, the virtual object is triggered to change from the first form to the second form.
Alternatively, the second output module 302 may be specifically configured to: acquiring a preset interaction point; and respectively controlling the target application and the virtual object to move towards the interaction point, and displaying the moving effect after the movement.
Alternatively, the second output module 302 may be specifically configured to: identifying or determining that the target application and the virtual object are located at an interaction point; amplifying an application icon of the target application at a preset position of the virtual object; and identifying or determining that the application icons are enlarged to a preset proportion, and displaying an application interface of the target application.
As described above, the interactive device of the application responds to the wake-up operation, outputs the virtual object, responds to the interactive operation, confirms or generates the dynamic effect of the target application based on the virtual object, and can respond to the interactive operation when the target application is started, output the dynamic effect of the virtual object for starting the target application, and enrich the interactive mode of the intelligent voice assistant and the user.
The embodiment of the application also provides an intelligent terminal, which comprises a memory and a processor, wherein the memory is stored with a computer program, and the computer program realizes the steps of the interaction method in any embodiment when being executed by the processor.
The embodiment of the application further provides a storage medium, on which a computer program is stored, which when executed by a processor, implements the steps of the interaction method in any of the above embodiments.
The embodiments of the intelligent terminal and the storage medium provided in the present application may include all technical features of any one of the embodiments of the interaction method, and the expansion and explanation contents of the description are substantially the same as those of each embodiment of the method, which are not repeated herein.
The present embodiments also provide a computer program product comprising computer program code which, when run on a computer, causes the computer to perform the method in the various possible implementations as above.
The embodiments also provide a chip including a memory for storing a computer program and a processor for calling and running the computer program from the memory, so that a device on which the chip is mounted performs the method in the above possible embodiments.
It can be understood that the above scenario is merely an example, and does not constitute a limitation on the application scenario of the technical solution provided in the embodiments of the present application, and the technical solution of the present application may also be applied to other scenarios. For example, as one of ordinary skill in the art can know, with the evolution of the system architecture and the appearance of new service scenarios, the technical solutions provided in the embodiments of the present application are equally applicable to similar technical problems.
The foregoing embodiment numbers of the present application are merely for describing, and do not represent advantages or disadvantages of the embodiments.
The steps in the method of the embodiment of the application can be sequentially adjusted, combined and deleted according to actual needs.
The units in the device of the embodiment of the application can be combined, divided and pruned according to actual needs.
In this application, the same or similar term concept, technical solution, and/or application scenario description will generally be described in detail only when first appearing, and when repeated later, for brevity, will not generally be repeated, and when understanding the content of the technical solution of the present application, etc., reference may be made to the previous related detailed description thereof for the same or similar term concept, technical solution, and/or application scenario description, etc., which are not described in detail later.
In this application, the descriptions of the embodiments are focused on, and the details or descriptions of one embodiment may be found in the related descriptions of other embodiments.
The technical features of the technical solutions of the present application may be arbitrarily combined, and for brevity of description, all possible combinations of the technical features in the above embodiments are not described, however, as long as there is no contradiction between the combinations of the technical features, they should be considered as the scope of the present application.
From the above description of the embodiments, it will be clear to those skilled in the art that the above-described embodiment method may be implemented by means of software plus a necessary general hardware platform, but of course may also be implemented by means of hardware, but in many cases the former is a preferred embodiment. Based on such understanding, the technical solution of the present application may be embodied essentially or in a part contributing to the prior art in the form of a software product stored in a storage medium (e.g. ROM/RAM, magnetic disk, optical disk) as above, including several instructions for causing a terminal device (which may be a mobile phone, a computer, a server, a controlled terminal, or a network device, etc.) to perform the method of each embodiment of the present application.
In the above embodiments, it may be implemented in whole or in part by software, hardware, firmware, or any combination thereof. When implemented in software, may be implemented in whole or in part in the form of a computer program product. The computer program product includes one or more computer instructions. When the computer program instructions are loaded and executed on a computer, the processes or functions in accordance with embodiments of the present application are produced in whole or in part. The computer may be a general purpose computer, a special purpose computer, a network of computers, or other programmable devices. The computer instructions may be stored in a storage medium or transmitted from one storage medium to another storage medium, for example, from one website, computer, server, or data center to another website, computer, server, or data center by a wired (e.g., coaxial cable, fiber optic, digital subscriber line), or wireless (e.g., infrared, wireless, microwave, etc.) means. The storage media may be any available media that can be accessed by a computer or a data storage device such as a server, data center, or the like that contains an integration of one or more available media. Usable media may be magnetic media (e.g., floppy disks, storage disks, magnetic tape), optical media (e.g., DVD), or semiconductor media (e.g., solid State Disk (SSD)), among others.
The foregoing description is only of the preferred embodiments of the present application, and is not intended to limit the scope of the claims, and all equivalent structures or equivalent processes using the descriptions and drawings of the present application, or direct or indirect application in other related technical fields are included in the scope of the claims of the present application.
Claims (10)
1. An interaction method, characterized by comprising the steps of:
s11: responding to the awakening operation and outputting a virtual object;
s12: and responding to the interactive operation, and confirming or generating the dynamic effect of the target application based on the virtual object.
2. The method according to claim 1, wherein the step S12 includes:
identifying or determining a target application corresponding to the interactive operation;
and confirming or generating dynamic effects of the target application based on the virtual object according to the position information of the target application.
3. The method according to claim 2, wherein said validating or generating a dynamic effect of the target application based on the virtual object according to the location information of the target application comprises:
establishing a moving path of the virtual object and the target application according to the position information of the target application;
after the virtual object is controlled to move to the target application according to the moving path, starting the target application;
And triggering the virtual object to perform morphological change when the target application is started.
4. A method according to claim 3, wherein triggering the virtual object to undergo a morphological change when the target application is open comprises:
acquiring a first form and a second form of the virtual object, wherein the first form corresponds to an initial state of the target application, and the second form corresponds to a starting state of the target application;
and triggering the virtual object to change from the first form to the second form when the target application is switched from the initial state to the starting state.
5. The method according to claim 2, wherein said validating or generating a dynamic effect of the target application based on the virtual object according to the location information of the target application comprises:
acquiring a preset interaction point;
and respectively controlling the target application and the virtual object to move towards the interaction point, and displaying the moving effect after the movement.
6. The method of claim 5, wherein controlling the movement of the target application and virtual object toward the interaction point and displaying the moved effects, respectively, comprises:
Identifying or determining that the target application and virtual object are located at the interaction point;
amplifying the application icon of the target application at a preset position of the virtual object;
and identifying or determining that the application icon is enlarged to a preset proportion, and displaying an application interface of the target application.
7. The method according to any one of claims 1 to 6, further comprising, after step S12:
and responding to the hiding operation aiming at the target application, and outputting the dynamic effect of hiding the target application by the virtual object.
8. The method of claim 7, wherein outputting the dynamic effect of the virtual object to hide the target application in response to the hide operation for the target application comprises at least one of:
responding to a minimizing operation aiming at a target application, and outputting dynamic effects of dragging and/or hiding the target application by the virtual object;
and responding to closing operation aiming at the target application, and outputting the virtual object desalination and/or the dynamic effect of closing the target application.
9. An intelligent terminal, characterized by comprising: memory, a processor, wherein the memory has stored thereon a computer program which, when executed by the processor, implements the steps of the interaction method of any of claims 1 to 8.
10. A storage medium having stored thereon a computer program which, when executed by a processor, implements the steps of the interaction method of any of claims 1 to 8.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310168271.7A CN116168701A (en) | 2023-02-24 | 2023-02-24 | Interaction method, intelligent terminal and storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310168271.7A CN116168701A (en) | 2023-02-24 | 2023-02-24 | Interaction method, intelligent terminal and storage medium |
Publications (1)
Publication Number | Publication Date |
---|---|
CN116168701A true CN116168701A (en) | 2023-05-26 |
Family
ID=86421743
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202310168271.7A Pending CN116168701A (en) | 2023-02-24 | 2023-02-24 | Interaction method, intelligent terminal and storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN116168701A (en) |
-
2023
- 2023-02-24 CN CN202310168271.7A patent/CN116168701A/en active Pending
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108572764B (en) | Character input control method and device and computer readable storage medium | |
CN107885448B (en) | Control method for application touch operation, mobile terminal and readable storage medium | |
CN113126844A (en) | Display method, terminal and storage medium | |
CN109710168B (en) | Screen touch method and device and computer readable storage medium | |
CN115469949A (en) | Information display method, intelligent terminal and storage medium | |
CN114443199A (en) | Interface processing method, intelligent terminal and storage medium | |
CN113805700A (en) | Voice control method, mobile terminal and storage medium | |
CN113419694A (en) | Processing method, mobile terminal and storage medium | |
CN116168701A (en) | Interaction method, intelligent terminal and storage medium | |
CN116324695A (en) | Interaction method, terminal and storage medium | |
CN109669594B (en) | Interaction control method, equipment and computer readable storage medium | |
CN116166363A (en) | Processing method, intelligent terminal and storage medium | |
CN117572998A (en) | Display method, intelligent terminal and storage medium | |
CN117032544A (en) | Processing method, intelligent terminal and storage medium | |
CN117742647A (en) | Audio adjusting method, intelligent terminal and storage medium | |
CN117193596A (en) | Display method, intelligent terminal and storage medium | |
CN117453088A (en) | Display method, intelligent terminal and storage medium | |
CN116225279A (en) | Adjustment method, intelligent terminal and storage medium | |
CN117707375A (en) | Interaction method, intelligent terminal and storage medium | |
CN117768866A (en) | Control method, intelligent terminal and storage medium | |
CN116366925A (en) | Processing method, intelligent terminal and storage medium | |
CN116185239A (en) | Data processing method, intelligent terminal and storage medium | |
CN116048704A (en) | Display method, intelligent terminal and storage medium | |
CN117135265A (en) | Display method, intelligent terminal and storage medium | |
CN116360659A (en) | Processing method, intelligent terminal and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication |