Detailed Description
It should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the scope of the invention.
In the following description, suffixes such as "module", "component", or "unit" for representing elements are used only for facilitating the description of the present invention, and have no specific meaning per se. Thus, "module," "component," or "unit" may be used in combination.
The wearable device provided by the embodiment of the invention comprises a mobile terminal such as an intelligent bracelet, an intelligent watch and an intelligent mobile phone. With the continuous development of screen technology, mobile terminals such as smart phones and the like can also be used as wearable devices due to the appearance of screen forms such as flexible screens, folding screens and the like. The wearable device provided in the embodiment of the invention can comprise: RF (Radio Frequency) unit, wiFi module, audio output unit, A/V (audio/video) input unit, sensor, display unit, user input unit, interface unit, memory, processor, and power supply.
In the following description, a wearable device will be taken as an example, please refer to fig. 1, which is a schematic hardware structure of a wearable device implementing various embodiments of the present invention, where the wearable device 100 may include: an RF (Radio Frequency) unit 101, a WiFi module 102, an audio output unit 103, an a/V (audio/video) input unit 104, a sensor 105, a display unit 106, a user input unit 107, an interface unit 108, a memory 109, a processor 110, and a power supply 111. Those skilled in the art will appreciate that the wearable device structure shown in fig. 1 does not constitute a limitation of the wearable device, and that the wearable device may include more or fewer components than shown, or certain components in combination, or a different arrangement of components.
The following describes the various components of the wearable device in detail with reference to fig. 1:
the radio frequency unit 101 may be used to send and receive information or send signals in a call process, specifically, the radio frequency unit 101 may send uplink information to the base station, or may send downlink information sent by the base station to the processor 110 of the wearable device to process the downlink information, where the downlink information sent by the base station to the radio frequency unit 101 may be generated according to the uplink information sent by the radio frequency unit 101, or may be actively pushed to the radio frequency unit 101 after detecting that the information of the wearable device is updated, for example, after detecting that the geographic position where the wearable device is located changes, the base station may send a notification of the change of the geographic position to the radio frequency unit 101 of the wearable device, after receiving the notification of the message, the radio frequency unit 101 may send the notification of the message to the processor 110 of the wearable device to process, and the processor 110 of the wearable device may control the notification of the message to be displayed on the display panel 1061 of the wearable device; typically, the radio frequency unit 101 includes, but is not limited to, an antenna, at least one amplifier, a transceiver, a coupler, a low noise amplifier, a duplexer, and the like. In addition, the radio frequency unit 101 may also communicate with a network and other devices through wireless communication, which may specifically include: through wireless communication with a server in a network system, for example, the wearable device can download file resources from the server through wireless communication, for example, an application program can be downloaded from the server, after the wearable device finishes downloading a certain application program, if the file resources corresponding to the application program in the server are updated, the server can push a message notification of the resource update to the wearable device through wireless communication so as to remind a user to update the application program. The wireless communication may use any communication standard or protocol, including but not limited to GSM (Global System of Mobile communication, global System for Mobile communications), GPRS (General Packet Radio Service ), CDMA2000 (Code Division Multiple Access, CDMA 2000), WCDMA (Wideband Code Division Multiple Access ), TD-SCDMA (Time Division-Synchronous Code Division Multiple Access, time Division synchronous code Division multiple Access), FDD-LTE (Frequency Division Duplexing-Long Term Evolution, frequency Division Duplex Long term evolution), and TDD-LTE (Time Division Duplexing-Long Term Evolution, time Division Duplex Long term evolution), etc.
In one embodiment, the wearable device 100 may access an existing communication network by inserting a SIM card.
In another embodiment, the wearable device 100 may access an existing communication network by setting an esim card (Embedded-SIM), and by adopting the esim card, the internal space of the wearable device may be saved and the thickness may be reduced.
It will be appreciated that although fig. 1 shows a radio frequency unit 101, it will be appreciated that the radio frequency unit 101 is not an essential component of a wearable device and may be omitted entirely as required within the scope of not changing the essence of the invention. The wearable device 100 may implement communication connection with other devices or communication networks through the wifi module 102 alone, which is not limited by the embodiment of the present invention.
WiFi belongs to a short-distance wireless transmission technology, and the wearable device can help a user to send and receive emails, browse webpages, access streaming media and the like through the WiFi module 102, so that wireless broadband Internet access is provided for the user. Although fig. 1 shows a WiFi module 102, it is understood that it does not belong to the necessary constitution of the wearable device, and can be omitted entirely as required within the scope of not changing the essence of the invention.
The audio output unit 103 may convert audio data received by the radio frequency unit 101 or the WiFi module 102 or stored in the memory 109 into an audio signal and output as sound when the wearable device 100 is in a call signal reception mode, a talk mode, a recording mode, a voice recognition mode, a broadcast reception mode, or the like. Also, the audio output unit 103 may also provide audio output (e.g., call signal reception sound, message reception sound, etc.) related to a specific function performed by the wearable device 100. The audio output unit 103 may include a speaker, a buzzer, and the like.
The a/V input unit 104 is used to receive an audio or video signal. The a/V input unit 104 may include a graphics processor (Graphics Processing Unit, GPU) 1041 and a microphone 1042, the graphics processor 1041 processing image data of still pictures or video obtained by an image capturing device (e.g., a camera) in a video capturing mode or an image capturing mode. The processed image frames may be displayed on the display unit 106. The image frames processed by the graphics processor 1041 may be stored in the memory 109 (or other storage medium) or transmitted via the radio frequency unit 101 or the WiFi module 102. The microphone 1042 can receive sound (audio data) via the microphone 1042 in a phone call mode, a recording mode, a voice recognition mode, and the like, and can process such sound into audio data. The processed audio (voice) data may be converted into a format output that can be transmitted to the mobile communication base station via the radio frequency unit 101 in the case of a telephone call mode. The microphone 1042 may implement various types of noise cancellation (or suppression) algorithms to cancel (or suppress) noise or interference generated in the course of receiving and transmitting the audio signal.
In one embodiment, the wearable device 100 includes one or more cameras, and by opening the cameras, capturing of images, photographing, video recording and other functions can be achieved, and the positions of the cameras can be set as required.
The wearable device 100 further comprises at least one sensor 105, such as a light sensor, a motion sensor, and other sensors. Specifically, the light sensor includes an ambient light sensor and a proximity sensor, wherein the ambient light sensor can adjust the brightness of the display panel 1061 according to the brightness of ambient light, and the proximity sensor can turn off the display panel 1061 and/or the backlight when the wearable device 100 moves to the ear. As one type of motion sensor, the accelerometer sensor can detect the acceleration in all directions (typically three axes), and can detect the gravity and direction when stationary, and can be used for applications for recognizing the gesture of a mobile phone (such as horizontal-vertical screen switching, related games, magnetometer gesture calibration), vibration recognition related functions (such as pedometer, knocking), and the like.
In one embodiment, the wearable device 100 further comprises a proximity sensor, by employing the proximity sensor, the wearable device is able to achieve non-contact manipulation, providing more modes of operation.
In one embodiment, the wearable device 100 further comprises a heart rate sensor, which when worn, enables detection of heart rate by being in close proximity to the user.
In one embodiment, the wearable device 100 may further include a fingerprint sensor, by reading a fingerprint, security verification or the like can be achieved.
The display unit 106 is used to display information input by a user or information provided to the user. The display unit 106 may include a display panel 1061, and the display panel 1061 may be configured in the form of a liquid crystal display (Liquid Crystal Display, LCD), an Organic Light-Emitting Diode (OLED), or the like.
In one embodiment, the display panel 1061 employs a flexible display screen, and the wearable device employing the flexible display screen is capable of bending when worn, thereby fitting more. Optionally, the flexible display screen may be an OLED screen body and a graphene screen body, and in other embodiments, the flexible display screen may also be other display materials, which is not limited to this embodiment.
In one embodiment, the display panel 1061 of the wearable device may take a rectangular shape for ease of wrapping when worn. In other embodiments, other approaches may be taken as well.
The user input unit 107 may be used to receive input numeric or character information and to generate key signal inputs related to user settings and function control of the wearable device. In particular, the user input unit 107 may include a touch panel 1071 and other input devices 1072. The touch panel 1071, also referred to as a touch screen, may collect touch operations thereon or thereabout by a user (e.g., operations of the user on the touch panel 1071 or thereabout by using any suitable object or accessory such as a finger, a stylus, etc.) and drive the corresponding connection device according to a predetermined program. The touch panel 1071 may include two parts of a touch detection device and a touch controller. The touch detection device detects the touch azimuth of a user, detects a signal brought by touch operation and transmits the signal to the touch controller; the touch controller receives touch information from the touch detection device, converts it into touch point coordinates, and sends the touch point coordinates to the processor 110, and can receive and execute commands sent from the processor 110. Further, the touch panel 1071 may be implemented in various types such as resistive, capacitive, infrared, and surface acoustic wave. The user input unit 107 may include other input devices 1072 in addition to the touch panel 1071. In particular, other input devices 1072 may include, but are not limited to, one or more of a physical keyboard, function keys (e.g., volume control keys, switch keys, etc.), a trackball, mouse, joystick, etc., as specifically not limited herein.
In one embodiment, the sides of the wearable device 100 may be provided with one or more buttons. The button can realize a plurality of modes such as short pressing, long pressing, rotation and the like, thereby realizing a plurality of operation effects. The number of the buttons can be multiple, and different buttons can be combined for use, so that multiple operation functions are realized.
Further, the touch panel 1071 may overlay the display panel 1061, and when the touch panel 1071 detects a touch operation thereon or thereabout, the touch panel 1071 is transferred to the processor 110 to determine the type of touch event, and then the processor 110 provides a corresponding visual output on the display panel 1061 according to the type of touch event. Although in fig. 1, the touch panel 1071 and the display panel 1061 are two independent components for implementing the input and output functions of the wearable device, in some embodiments, the touch panel 1071 may be integrated with the display panel 1061 to implement the input and output functions of the wearable device, which is not limited herein. For example, when a message notification of a certain application is received through the rf unit 101, the processor 110 may control the message notification to be displayed in a certain preset area of the display panel 1061, where the preset area corresponds to a certain area of the touch panel 1071, and may control the message notification displayed in the corresponding area on the display panel 1061 by performing a touch operation on the certain area of the touch panel 1071.
The interface unit 108 serves as an interface through which at least one external device can be connected with the wearable apparatus 100. For example, the external devices may include a wired or wireless headset port, an external power (or battery charger) port, a wired or wireless data port, a memory card port, a port for connecting a device having an identification module, an audio input/output (I/O) port, a video I/O port, an earphone port, and the like. The interface unit 108 may be used to receive input (e.g., data information, power, etc.) from an external device and transmit the received input to one or more elements within the wearable apparatus 100 or may be used to transmit data between the wearable apparatus 100 and the external device.
In one embodiment, the interface unit 108 of the wearable device 100 adopts a contact structure, and is connected with other corresponding devices through the contact, so as to realize functions of charging, connection and the like. The contact can also be waterproof.
Memory 109 may be used to store software programs as well as various data. The memory 109 may mainly include a storage program area that may store an operating system, application programs required for at least one function (such as a sound playing function, an image playing function, etc.), and a storage data area; the storage data area may store data (such as audio data, phonebook, etc.) created according to the use of the handset, etc. In addition, memory 109 may include high-speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other volatile solid-state storage device.
The processor 110 is a control center of the wearable device, connects various parts of the entire wearable device with various interfaces and lines, performs various functions of the wearable device and processes data by running or executing software programs and/or modules stored in the memory 109, and invoking data stored in the memory 109, thereby performing overall monitoring of the wearable device. Processor 110 may include one or more processing units; preferably, the processor 110 may integrate an application processor that primarily handles operating systems, user interfaces, applications, etc., with a modem processor that primarily handles wireless communications. It will be appreciated that the modem processor described above may not be integrated into the processor 110.
The wearable device 100 may further include a power source 111 (such as a battery) for powering the various components, and preferably, the power source 111 may be logically connected to the processor 110 through a power management system, so as to perform functions of managing charging, discharging, and power consumption management through the power management system.
Although not shown in fig. 1, the wearable device 100 may further include a bluetooth module or the like, which is not described herein. The wearable device 100 can be connected with other terminal devices through bluetooth to realize communication and information interaction.
Fig. 2 to fig. 4 are schematic structural diagrams of a wearable device according to an embodiment of the present invention. The wearable device comprises a flexible screen. When the wearable device is unfolded, the flexible screen is in a strip shape; when the wearable device is in a wearing state, the flexible screen is bent to be annular. Fig. 2 and 3 show schematic structural diagrams of the wearable device screen when unfolded, and fig. 4 shows schematic structural diagrams of the wearable device screen when bent.
Based on the above embodiments, it can be seen that if the device is a wristwatch, a bracelet, or a wearable device, the screen of the device may not cover the watchband area of the device, or may cover the watchband area of the device. In this embodiment, the device may be a wristwatch, a bracelet, or a wearable device, and the device includes a screen and a connection portion. The screen may be a flexible screen and the connection may be a wristband. Alternatively, the screen of the device or the display area of the screen may be partially or fully overlaid on the wristband of the device. Fig. 5 is a schematic hardware diagram of an implementation manner of a wearable device according to an embodiment of the present application, where a screen of the device extends to two sides, and a part of the screen is covered on a watchband of the device. In other embodiments, the screen of the device may also be entirely covered on the watchband of the device, which is not limited to this embodiment.
Example 1
Fig. 6 is a flowchart of a first embodiment of a video recording interaction control method of the present invention. A video recording interactive control method, the method comprising:
s1, acquiring a video recording state and a sharing requirement on wearable equipment;
s2, analyzing the sharing requirement, and displaying a sharing object in an associated area of the video preview area;
s3, identifying and marking a shooting object with a first identification in the video preview area in a preset first sharing time;
s4, marking the sharing object with a second identification in the association area in a preset second sharing time.
In this embodiment, first, a video recording state and a sharing requirement on a wearable device are obtained; then, analyzing the sharing requirement, and displaying a sharing object in an associated area of the video preview area; then, in a preset first sharing time, identifying and marking a shooting object with a first identification in the video preview area; and finally, marking the sharing object with a second identification in the association area in a preset second sharing time.
In view of the fact that in the prior art, particularly when shooting previews are performed on a wearable device, the display area of the wearable device is relatively narrow, and because the video recording previewing area has a specific aspect ratio, even if a relatively wide horizontal display space exists in the display area of the wearable device, the video recording previewing area is relatively narrow due to the shortage of the vertical display space, if a user needs to perform related processing on the recorded video in the video recording previewing process, a related operation related to the video recording state or the video recording content is displayed in the related area of the video recording previewing area, for example, when the user records a small video, the user can display a commonly used contact in the instant messaging application program or the instant messaging application program in the upper end area and the lower end area of the recording previewing area, or identify the video recording content, and when the commonly used recording object of the user is contained in the video recording content, a target contact to be selected for historically transmitting the recording object is displayed in the related area of the video recording previewing area. Therefore, a more convenient video recording interactive operation scheme is realized, so that when a user records videos, the display space of the wearable device is fully utilized, the shot videos are shared more conveniently, the inconvenience of multiple switching in a narrow display area is avoided, the operation efficiency is improved, and the user experience is enhanced.
The video recording method and the video recording device have the advantages that video recording states and sharing requirements on the wearable equipment are obtained; then, analyzing the sharing requirement, and displaying a sharing object in an associated area of the video preview area; then, in a preset first sharing time, identifying and marking a shooting object with a first identification in the video preview area; and finally, marking the sharing object with a second identification in the association area in a preset second sharing time. The humanized video recording interaction control scheme is realized, so that when a user records a video through the wearable device, the display space of the wearable device is fully utilized, the video sharing operation efficiency is improved, and the user experience is enhanced.
Example two
Fig. 7 is a flowchart of a second embodiment of a video recording interaction control method according to the present invention, based on the above embodiment, the obtaining a video recording state on a wearable device, and determining a corresponding sharing requirement according to the video recording state, includes:
s11, acquiring the wearing state of the wearing equipment;
s12, determining the video recording direction according to the wearing state and the position of the camera shooting component of the wearing equipment.
In this embodiment, first, a wearing state of the wearing device is obtained; and then, determining the video recording position according to the wearing state and the position of the camera component of the wearing equipment.
Optionally, acquiring a detection image according to an image sensor of the wearable device, and then acquiring a wearing state of the wearable device according to image features (features such as azimuth, distance, horizontal height and the like) of the detection image;
optionally, the video recording position is determined according to the wearing state and the position of the camera component of the wearing device, and similarly, as described above for an example, the recording position of the camera component is determined according to the features of the position, the distance, the horizontal height and the like of the detected image.
The beneficial effects of the embodiment are that the wearing state of the wearing equipment is obtained; and then, determining the video recording position according to the wearing state and the position of the camera component of the wearing equipment. The video recording interaction control scheme is more humanized, so that when a user records a video through the wearable device, the display space of the wearable device is fully utilized, the operation efficiency of video sharing is improved, and the user experience is enhanced.
Example III
Fig. 8 is a flowchart of a third embodiment of a video recording interaction control method according to the present invention, based on the above embodiment, the method for obtaining a video recording state on a wearable device, and determining a corresponding sharing requirement according to the video recording state, further includes:
s13, under the video recording direction, acquiring video recording data through the camera component and acquiring the sharing requirement;
s14, dividing a video preview area in a display area of the wearable device by combining the wearing state of the wearable device and the position of the camera component of the wearable device.
In this embodiment, first, under the video recording direction, video recording data is obtained through the camera component, and the sharing requirement is obtained; then, dividing a video preview area in a display area of the wearable device by combining the wearing state of the wearable device and the position of the camera component of the wearable device.
Optionally, in this embodiment, in the video recording direction, the video recording data is obtained through the camera component, and the sharing requirement is obtained, where the sharing requirement includes a sharing way and a sharing application program, and specifically, the sharing way includes a sharing network link mode, and the sharing application program includes a sharing application program type and a function module under the application program;
Optionally, in this embodiment, in combination with the wearing state of the wearable device and the position of the camera component of the wearable device, a video preview area is obtained by dividing in a display area of the wearable device, where, in order to facilitate the user to view and operate, particularly, according to the number and arrangement states of the shared application programs, the video preview area is obtained by dividing in the display area of the wearable device, so that the user can still clearly view the shared application programs or the functional modules under the application programs while viewing the video preview area.
The video recording method has the advantages that video recording data are obtained through the camera component under the video recording direction, and the sharing requirement is obtained; then, dividing a video preview area in a display area of the wearable device by combining the wearing state of the wearable device and the position of the camera component of the wearable device. The video recording interaction control scheme is more humanized, so that when a user records a video through the wearable device, the display space of the wearable device is fully utilized, the operation efficiency of video sharing is improved, and the user experience is enhanced.
Example IV
Fig. 9 is a flowchart of a fourth embodiment of a video recording interaction control method according to the present invention, based on the above embodiment, the analyzing the sharing requirement, and displaying the sharing object corresponding to the sharing requirement in an associated area of a video preview area, includes:
s21, analyzing the sharing requirement, wherein the sharing requirement comprises a sharing type requirement and a sharing time requirement;
and S22, displaying the video preview image in the video preview area, and determining an association area associated with the video preview area in the display area.
In this embodiment, first, the sharing requirement is parsed, where the sharing requirement includes a sharing type requirement and a sharing time requirement; then, a video preview image is displayed within the video preview area while an associated area associated with the video preview area is determined within the display area.
Optionally, in this embodiment, the sharing requirement is parsed, where the sharing requirement includes a sharing type requirement and a sharing time requirement, the specific sharing type requirement includes a type of a shared video and a type of a shared video content, and the sharing time requirement includes a video length of the shared video;
Optionally, a video preview image is displayed in the video preview area, and at the same time, an associated area associated with the video preview area is determined in the display area, and, as described above for example, in order to facilitate the user to view and operate, in particular, according to the number and arrangement states of the shared application programs, the video preview area is divided into the display area of the wearable device, so that the user can still clearly view the associated area while viewing the video preview area.
The beneficial effects of the embodiment are that the sharing requirement comprises a sharing type requirement and a sharing time requirement through analysis; then, a video preview image is displayed within the video preview area while an associated area associated with the video preview area is determined within the display area. The video recording interaction control scheme is more humanized, so that when a user records a video through the wearable device, the display space of the wearable device is fully utilized, the operation efficiency of video sharing is improved, and the user experience is enhanced.
Example five
Fig. 10 is a flowchart of a fifth embodiment of a video recording interaction control method according to the present invention, based on the above embodiment, the analyzing the sharing requirement, and displaying the sharing object corresponding to the sharing requirement in an associated area of a video preview area, further includes:
S23, traversing an application program in the wearable device, and extracting a sharing object related to the application program, wherein the sharing object comprises a contact object, a photo object, a text object and a video object;
and S24, displaying the sharing object in the association area of the video preview area.
In this embodiment, firstly, traversing an application program in the wearable device, and extracting a sharing object related to the application program, where the sharing object includes a contact object, a photo object, a text object and a video object; and then displaying the sharing object in the association area of the video preview area.
Optionally, traversing an application program in the wearable device, and extracting a sharing object related to the application program, where the sharing object includes a contact object, a photo object, a text object and a video object, and specifically, the sharing object includes an avatar of a contact of the instant messaging application program, a photo character or a background in a recent album, chat text content of the recent instant messaging application program, a video shooting character in the recent album, and the like;
optionally, displaying the sharing object in the associated area of the video preview area according to display modes such as types, affiliated application programs and the like.
The method has the advantages that sharing objects related to the application program are extracted by traversing the application program in the wearable device, wherein the sharing objects comprise contact objects, photo objects, text objects and video objects; and then displaying the sharing object in the association area of the video preview area. The video recording interaction control scheme is more humanized, so that when a user records a video through the wearable device, the display space of the wearable device is fully utilized, the operation efficiency of video sharing is improved, and the user experience is enhanced.
Example six
Fig. 11 is a flowchart of a sixth embodiment of a video recording interaction control method according to the present invention, based on the above embodiment, the identifying and marking, with a first identification mark, a shooting object associated with the sharing object in the video preview area within a preset first sharing time includes:
s31, presetting a first sharing time corresponding to the marking operation of the shooting object;
s32, identifying the shooting object in the video preview area, and marking the shooting object with a first identification.
In this embodiment, first, a first sharing time corresponding to a marking operation of the shooting object is preset; then, in the video preview area, the shooting object is identified, and in the first sharing time, the shooting object is marked with a first identification.
Optionally, in this embodiment, a first sharing time corresponding to the marking operation of the shooting object is preset, in the video recording process, in a time when the shooting object exists in the video preview area, marking is continuously performed, and if a continuous duration or an intermittent accumulated duration reaches the first sharing time, marking is canceled, so that the preview effect of the video recording is not affected by the long-time marking;
optionally, in the video preview area, the shooting object is identified, and in the first sharing time, the shooting object is continuously marked with a first identifier, which can be understood that one or more shooting objects can be shot at the same time, and the corresponding first identifier also includes multiple types of identifiers.
The method has the advantages that the first sharing time corresponding to the marking operation of the shooting object is preset; then, in the video preview area, the shooting object is identified, and in the first sharing time, the shooting object is marked with a first identification. The video recording interaction control scheme is more humanized, so that when a user records a video through the wearable device, the display space of the wearable device is fully utilized, the operation efficiency of video sharing is improved, and the user experience is enhanced.
Example seven
Fig. 12 is a flowchart of a seventh embodiment of a video recording interaction control method according to the present invention, based on the foregoing embodiment, where in the association area, a sharing object associated with the shooting object is marked with a second identifier in a preset second sharing time, including:
s41, presetting a second sharing time corresponding to the marking operation of the sharing object;
s42, identifying the sharing object in the association area, and marking the sharing object with a second identification.
In this embodiment, first, a second sharing time corresponding to a marking operation of the sharing object is preset; and then, identifying the sharing object in the association area, and marking the sharing object with a second identification in the second sharing time.
Optionally, a second sharing time corresponding to the marking operation of the sharing object is preset, and similarly, as described in the above example, in the video recording process, in the time when the shooting object exists in the video preview area, the continuous duration or the intermittent accumulated duration continuously marks the analysis object, and when the continuous duration or the intermittent accumulated duration reaches the second sharing time, the marking is cancelled, so as to avoid the influence of the long-time marking on the display effect of the associated area;
Optionally, in the association area, the sharing object is identified, and in the second sharing time, the sharing object is marked with a second identifier, and similarly, as described above for example, at the same time, one or more sharing objects corresponding to the shooting object may be corresponding to the second identifier, where the second identifier also includes multiple types of identifiers.
The method has the advantages that the second sharing time corresponding to the marking operation of the sharing object is preset; and then, identifying the sharing object in the association area, and marking the sharing object with a second identification in the second sharing time. The video recording interaction control scheme is more humanized, so that when a user records a video through the wearable device, the display space of the wearable device is fully utilized, the operation efficiency of video sharing is improved, and the user experience is enhanced.
Example eight
Fig. 13 is a flowchart of an eighth embodiment of a video recording interaction control method according to the present invention, based on the foregoing embodiment, where in the association area, a sharing object associated with the shooting object is marked with a second identifier in a preset second sharing time, further includes:
S43, when the sharing object is marked, determining corresponding sharing duration;
s44, sharing the video containing the marked sharing duration of the shooting object to the sharing object.
In this embodiment, first, when the sharing object is marked, a corresponding sharing duration is determined; and then, sharing the video containing the marked sharing duration of the shooting object to the sharing object.
Optionally, when the sharing object is marked, determining a corresponding sharing duration, where the sharing duration is determined by the sharing time requirement;
optionally, the video including the marked sharing duration of the shooting object is shared to the sharing object, and it can be understood that, under the above sharing time requirement, if the shooting object disappears, the video in the period of disappearance of the shooting object is correspondingly cut and removed.
The method has the beneficial effects that when the sharing object is marked, corresponding sharing duration is determined through judgment; and then, sharing the video containing the marked sharing duration of the shooting object to the sharing object. The video recording interaction control scheme is more humanized, so that when a user records a video through the wearable device, the display space of the wearable device is fully utilized, the operation efficiency of video sharing is improved, and the user experience is enhanced.
Example nine
Based on the above embodiment, the present invention further provides a video recording interaction control device, which includes:
a memory, a processor, and a computer program stored on the memory and executable on the processor;
the computer program implementing the steps of the method according to any of the preceding claims when executed by the processor.
Specifically, in this embodiment, first, a video recording state on a wearable device is obtained, and a corresponding sharing requirement is determined according to the video recording state; then, analyzing the sharing requirement, and displaying the sharing requirement in a sharing object corresponding to the sharing requirement in an associated area of a video preview area; then, in a preset first sharing time, identifying and marking a shooting object associated with the sharing object by a first identification in the video preview area; and finally, marking the sharing object associated with the shooting object in the association area by a second identification in a preset second sharing time.
In view of the fact that in the prior art, particularly when shooting previews are performed on a wearable device, the display area of the wearable device is relatively narrow, and because the video recording previewing area has a specific aspect ratio, even if a relatively wide horizontal display space exists in the display area of the wearable device, the video recording previewing area is relatively narrow due to the shortage of the vertical display space, if a user needs to perform related processing on the recorded video in the video recording previewing process, a related operation related to the video recording state or the video recording content is displayed in the related area of the video recording previewing area, for example, when the user records a small video, the user can display a commonly used contact in the instant messaging application program or the instant messaging application program in the upper end area and the lower end area of the recording previewing area, or identify the video recording content, and when the commonly used recording object of the user is contained in the video recording content, a target contact to be selected for historically transmitting the recording object is displayed in the related area of the video recording previewing area. Therefore, a more convenient video recording interactive operation scheme is realized, so that when a user records videos, the display space of the wearable device is fully utilized, the shot videos are shared more conveniently, the inconvenience of multiple switching in a narrow display area is avoided, the operation efficiency is improved, and the user experience is enhanced.
The method has the advantages that the video recording state on the wearable equipment is obtained, and the corresponding sharing requirement is determined according to the video recording state; then, analyzing the sharing requirement, and displaying the sharing requirement in a sharing object corresponding to the sharing requirement in an associated area of a video preview area; then, in a preset first sharing time, identifying and marking a shooting object associated with the sharing object by a first identification in the video preview area; and finally, marking the sharing object associated with the shooting object in the association area by a second identification in a preset second sharing time. The humanized video recording interaction control scheme is realized, so that when a user records a video through the wearable device, the display space of the wearable device is fully utilized, the video sharing operation efficiency is improved, and the user experience is enhanced.
Examples ten
Based on the above embodiments, the present invention also proposes a computer readable storage medium having a bitmap processing program stored thereon, which when executed by a processor implements the steps of the bitmap processing method according to any one of the above.
By implementing the bitmap processing method, the bitmap processing device and the computer readable storage medium, the video recording state on the wearable device is obtained, and the corresponding sharing requirement is determined according to the video recording state; then, analyzing the sharing requirement, and displaying the sharing requirement in a sharing object corresponding to the sharing requirement in an associated area of a video preview area; then, in a preset first sharing time, identifying and marking a shooting object associated with the sharing object by a first identification in the video preview area; and finally, marking the sharing object associated with the shooting object in the association area by a second identification in a preset second sharing time. The humanized video recording interaction control scheme is realized, so that when a user records a video through the wearable device, the display space of the wearable device is fully utilized, the video sharing operation efficiency is improved, and the user experience is enhanced.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising one … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element.
The foregoing embodiment numbers of the present invention are merely for the purpose of description, and do not represent the advantages or disadvantages of the embodiments.
From the above description of the embodiments, it will be clear to those skilled in the art that the above-described embodiment method may be implemented by means of software plus a necessary general hardware platform, but of course may also be implemented by means of hardware, but in many cases the former is a preferred embodiment. Based on such understanding, the technical solution of the present invention may be embodied essentially or in a part contributing to the prior art in the form of a software product stored in a storage medium (e.g. ROM/RAM, magnetic disk, optical disk) comprising instructions for causing a terminal (which may be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.) to perform the method according to the embodiments of the present invention.
The embodiments of the present invention have been described above with reference to the accompanying drawings, but the present invention is not limited to the above-described embodiments, which are merely illustrative and not restrictive, and many forms may be made by those having ordinary skill in the art without departing from the spirit of the present invention and the scope of the claims, which are to be protected by the present invention.