CN110620875B - Screenshot control method, equipment and computer readable storage medium in video shooting process - Google Patents

Screenshot control method, equipment and computer readable storage medium in video shooting process Download PDF

Info

Publication number
CN110620875B
CN110620875B CN201910917201.0A CN201910917201A CN110620875B CN 110620875 B CN110620875 B CN 110620875B CN 201910917201 A CN201910917201 A CN 201910917201A CN 110620875 B CN110620875 B CN 110620875B
Authority
CN
China
Prior art keywords
screenshot
video
area
shooting
preview area
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910917201.0A
Other languages
Chinese (zh)
Other versions
CN110620875A (en
Inventor
邵雪纯
杨亮
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nubia Technology Co Ltd
Original Assignee
Nubia Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nubia Technology Co Ltd filed Critical Nubia Technology Co Ltd
Priority to CN201910917201.0A priority Critical patent/CN110620875B/en
Publication of CN110620875A publication Critical patent/CN110620875A/en
Application granted granted Critical
Publication of CN110620875B publication Critical patent/CN110620875B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04BTRANSMISSION
    • H04B1/00Details of transmission systems, not covered by a single one of groups H04B3/00 - H04B13/00; Details of transmission systems not characterised by the medium used for transmission
    • H04B1/38Transceivers, i.e. devices in which transmitter and receiver form a structural unit and in which at least one part is used for functions of transmitting and receiving
    • H04B1/3827Portable transceivers
    • H04B1/385Transceivers carried on the body, e.g. in helmets
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/62Control of parameters via user interfaces
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/695Control of camera direction for changing a field of view, e.g. pan, tilt or based on tracking of objects

Landscapes

  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Human Computer Interaction (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The application discloses a screenshot control method, equipment and a computer-readable storage medium in a video shooting process, wherein the method comprises the following steps: dividing a video preview area in a display area in the sight range, and dividing a screenshot preview area in a display area outside the sight range; and in the shooting process, acquiring a screenshot instruction of a control area of the video area, generating a video screenshot according to the screenshot instruction, and displaying the video screenshot in the screenshot preview area in a distributed manner. The screenshot control scheme in the humanized shooting process is realized, so that when a user records a video through the wearable device, the display space of the wearable device is fully utilized, historical screenshot image viewing is conveniently carried out, inconvenience of repeated switching in a narrow display area is avoided, the operation efficiency is improved, and the user experience is enhanced.

Description

Screenshot control method, equipment and computer readable storage medium in video shooting process
Technical Field
The present application relates to the field of mobile communications, and in particular, to a method and apparatus for controlling screenshot in a video capturing process, and a computer readable storage medium.
Background
In the prior art, with the rapid development of intelligent terminal equipment, wearable equipment different from conventional smart phones, such as wearable equipment like a smart watch or a smart bracelet, appears. Because wearing formula equipment compares in traditional smart mobile phone, its particularities such as software, hardware environment, operation mode and operation environment, if the scheme of controlling of traditional smart mobile phone is transferred to wearing formula equipment, can bring inconvenience, user experience bad for user's operation.
Disclosure of Invention
In order to solve the technical defects in the prior art, the invention provides a screenshot control method in a video shooting process, which comprises the following steps:
acquiring a wearing state and a shooting state of a wearing device, determining a sight line range of the wearing device according to the wearing state, and determining a shooting azimuth of the wearing device according to the shooting state;
dividing a video preview area in a display area in the sight range, and dividing a screenshot preview area in a display area outside the sight range;
according to the dynamic change state of the shooting azimuth, the video preview area and the screenshot preview area are adjusted in real time;
And in the shooting process, acquiring a screenshot instruction of a control area of the video area, generating a video screenshot according to the screenshot instruction, and displaying the video screenshot in the screenshot preview area in a distributed manner.
Optionally, the acquiring the wearing state and the shooting state of the wearable device, determining the line of sight range of the wearable device according to the wearing state, and determining the shooting azimuth of the wearable device according to the shooting state includes:
acquiring the wearing state according to a pressure sensing component of the wearing device, wherein the wearing state comprises a wearing position and an operating area;
and determining the shooting state according to the wearing position of the wearing equipment and the shooting component of the wearing equipment, wherein the shooting state comprises a shooting azimuth and the shooting component adopted.
Optionally, the dividing the display area within the line of sight to obtain a video preview area, and the dividing the display area outside the line of sight to obtain a screenshot preview area includes:
determining the line of sight according to the wearing position;
and dividing the display area of the wearable equipment according to the shooting state in the sight range to obtain the video preview area.
Optionally, the dividing the display area within the line of sight to obtain a video preview area, and the dividing the display area outside the line of sight to obtain a screenshot preview area further includes:
shielding a touch signal in a display area outside the sight line range;
and dividing the screen capturing preview area in the display range of the shielding touch signal, wherein the initial area of the screen capturing preview area is the central area of the display area outside the sight line range.
Optionally, the adjusting the video preview area and the screenshot preview area in real time according to the dynamic change state of the shooting direction includes:
acquiring a dynamic change state of the shooting azimuth, wherein the dynamic change state comprises a change parameter of a shooting angle and a change parameter of an adopted shooting assembly;
and determining a corresponding adjustment reference value according to the dynamic change state, wherein the shooting angle is used as a moving direction reference value of the video preview area and the screenshot preview area.
Optionally, the adjusting the video preview area and the screenshot preview area in real time according to the dynamic change state of the shooting direction further includes:
Respectively moving the video preview area and the screenshot preview area to target positions according to the moving direction reference value;
and respectively taking the corresponding target positions as display centers, and adjusting the region ranges of the video preview region and the screenshot preview region.
Optionally, in the shooting process, acquiring a screenshot instruction of a control area of the video area, generating a video screenshot according to the screenshot instruction, and arranging and displaying the video screenshot in the screenshot preview area, where the method includes:
acquiring a screenshot instruction in the video area in the shooting process, wherein the screenshot instruction comprises a screenshot instruction generated by a touch signal;
and determining a target video frame according to the trigger time of the screenshot instruction, and synthesizing multi-frame images in the front and rear ranges of the target video frame to obtain the video screenshot.
Optionally, in the shooting process, a screenshot instruction of a control area of the video area is obtained, a video screenshot is generated according to the screenshot instruction, and the video screenshot is arranged and displayed in the screenshot preview area, and the method further includes:
moving the first video screenshot to a display center of the screenshot preview area;
And sequentially arranging and displaying the subsequent video screenshots on the two side areas of the display center of the screenshot preview area according to a time axis.
The invention also provides screenshot control equipment in the video shooting process, which comprises the following steps:
a memory, a processor, and a computer program stored on the memory and executable on the processor;
the computer program implementing the steps of the method according to any of the preceding claims when executed by the processor.
The invention also provides a computer readable storage medium, wherein the computer readable storage medium stores a screenshot control program in the video shooting process, and the screenshot control program in the video shooting process realizes the steps of the screenshot control method in the video shooting process when being executed by a processor.
The method has the advantages that the wearing state and the shooting state of the wearing equipment are obtained, the sight range of the wearing equipment is determined according to the wearing state, and the shooting direction of the wearing equipment is determined according to the shooting state; then, dividing a video preview area in a display area in the sight line range, and dividing a screenshot preview area in a display area outside the sight line range; then, according to the dynamic change state of the shooting azimuth, the video preview area and the screenshot preview area are adjusted in real time; and finally, in the shooting process, acquiring a screenshot instruction of a control area of the video area, generating a video screenshot according to the screenshot instruction, and displaying the video screenshot in the screenshot preview area in an arrangement mode. The screenshot control scheme in the humanized shooting process is realized, so that when a user records a video through the wearable device, the display space of the wearable device is fully utilized, historical screenshot image viewing is conveniently carried out, inconvenience of repeated switching in a narrow display area is avoided, the operation efficiency is improved, and the user experience is enhanced.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the application and together with the description, serve to explain the principles of the application.
In order to more clearly illustrate the embodiments of the application or the technical solutions of the prior art, the drawings which are used in the description of the embodiments or the prior art will be briefly described, and it will be obvious to a person skilled in the art that other drawings can be obtained from these drawings without inventive effort.
Fig. 1 is a schematic hardware structure of an implementation manner of a wearable device according to an embodiment of the present application;
fig. 2 is a hardware schematic diagram of an implementation manner of a wearable device according to an embodiment of the present application;
fig. 3 is a hardware schematic of an implementation manner of a wearable device according to an embodiment of the present application;
fig. 4 is a hardware schematic of an implementation manner of a wearable device according to an embodiment of the present application;
fig. 5 is a hardware schematic of an implementation manner of a wearable device according to an embodiment of the present application;
FIG. 6 is a flowchart of a first embodiment of a screenshot control method in a video capture process of the present application;
FIG. 7 is a flowchart of a second embodiment of a screenshot control method in a video capture process of the present invention;
FIG. 8 is a flowchart of a third embodiment of a screenshot control method in a video capture process of the present invention;
FIG. 9 is a flowchart of a fourth embodiment of a screenshot control method in a video capture process of the present invention;
FIG. 10 is a flowchart of a fifth embodiment of a screenshot control method in a video capture process of the present invention;
FIG. 11 is a flowchart of a sixth embodiment of a screenshot control method in a video capture process of the present invention;
FIG. 12 is a flowchart of a seventh embodiment of a screenshot control method in a video capture process of the present invention;
fig. 13 is a flowchart of an eighth embodiment of a screenshot control method in a video capture process according to the present invention.
Detailed Description
It should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the scope of the invention.
In the following description, suffixes such as "module", "component", or "unit" for representing elements are used only for facilitating the description of the present invention, and have no specific meaning per se. Thus, "module," "component," or "unit" may be used in combination.
The wearable device provided by the embodiment of the invention comprises a mobile terminal such as an intelligent bracelet, an intelligent watch and an intelligent mobile phone. With the continuous development of screen technology, mobile terminals such as smart phones and the like can also be used as wearable devices due to the appearance of screen forms such as flexible screens, folding screens and the like. The wearable device provided in the embodiment of the invention can comprise: RF (Radio Frequency) unit, wiFi module, audio output unit, A/V (audio/video) input unit, sensor, display unit, user input unit, interface unit, memory, processor, and power supply.
In the following description, a wearable device will be taken as an example, please refer to fig. 1, which is a schematic hardware structure of a wearable device implementing various embodiments of the present invention, where the wearable device 100 may include: an RF (Radio Frequency) unit 101, a WiFi module 102, an audio output unit 103, an a/V (audio/video) input unit 104, a sensor 105, a display unit 106, a user input unit 107, an interface unit 108, a memory 109, a processor 110, and a power supply 111. Those skilled in the art will appreciate that the wearable device structure shown in fig. 1 does not constitute a limitation of the wearable device, and that the wearable device may include more or fewer components than shown, or certain components in combination, or a different arrangement of components.
The following describes the various components of the wearable device in detail with reference to fig. 1:
the radio frequency unit 101 may be used to send and receive information or send signals in a call process, specifically, the radio frequency unit 101 may send uplink information to the base station, or may send downlink information sent by the base station to the processor 110 of the wearable device to process the downlink information, where the downlink information sent by the base station to the radio frequency unit 101 may be generated according to the uplink information sent by the radio frequency unit 101, or may be actively pushed to the radio frequency unit 101 after detecting that the information of the wearable device is updated, for example, after detecting that the geographic position where the wearable device is located changes, the base station may send a notification of the change of the geographic position to the radio frequency unit 101 of the wearable device, after receiving the notification of the message, the radio frequency unit 101 may send the notification of the message to the processor 110 of the wearable device to process, and the processor 110 of the wearable device may control the notification of the message to be displayed on the display panel 1061 of the wearable device; typically, the radio frequency unit 101 includes, but is not limited to, an antenna, at least one amplifier, a transceiver, a coupler, a low noise amplifier, a duplexer, and the like. In addition, the radio frequency unit 101 may also communicate with a network and other devices through wireless communication, which may specifically include: through wireless communication with a server in a network system, for example, the wearable device can download file resources from the server through wireless communication, for example, an application program can be downloaded from the server, after the wearable device finishes downloading a certain application program, if the file resources corresponding to the application program in the server are updated, the server can push a message notification of the resource update to the wearable device through wireless communication so as to remind a user to update the application program. The wireless communication may use any communication standard or protocol, including but not limited to GSM (Global System of Mobile communication, global System for Mobile communications), GPRS (General Packet Radio Service ), CDMA2000 (Code Division Multiple Access, CDMA 2000), WCDMA (Wideband Code Division Multiple Access ), TD-SCDMA (Time Division-Synchronous Code Division Multiple Access, time Division synchronous code Division multiple Access), FDD-LTE (Frequency Division Duplexing-Long Term Evolution, frequency Division Duplex Long term evolution), and TDD-LTE (Time Division Duplexing-Long Term Evolution, time Division Duplex Long term evolution), etc.
In one embodiment, the wearable device 100 may access an existing communication network by inserting a SIM card.
In another embodiment, the wearable device 100 may access an existing communication network by setting an esim card (Embedded-SIM), and by adopting the esim card, the internal space of the wearable device may be saved and the thickness may be reduced.
It will be appreciated that although fig. 1 shows a radio frequency unit 101, it will be appreciated that the radio frequency unit 101 is not an essential component of a wearable device and may be omitted entirely as required within the scope of not changing the essence of the invention. The wearable device 100 may implement communication connection with other devices or communication networks through the wifi module 102 alone, which is not limited by the embodiment of the present invention.
WiFi belongs to a short-distance wireless transmission technology, and the wearable device can help a user to send and receive emails, browse webpages, access streaming media and the like through the WiFi module 102, so that wireless broadband Internet access is provided for the user. Although fig. 1 shows a WiFi module 102, it is understood that it does not belong to the necessary constitution of the wearable device, and can be omitted entirely as required within the scope of not changing the essence of the invention.
The audio output unit 103 may convert audio data received by the radio frequency unit 101 or the WiFi module 102 or stored in the memory 109 into an audio signal and output as sound when the wearable device 100 is in a call signal reception mode, a talk mode, a recording mode, a voice recognition mode, a broadcast reception mode, or the like. Also, the audio output unit 103 may also provide audio output (e.g., call signal reception sound, message reception sound, etc.) related to a specific function performed by the wearable device 100. The audio output unit 103 may include a speaker, a buzzer, and the like.
The a/V input unit 104 is used to receive an audio or video signal. The a/V input unit 104 may include a graphics processor (Graphics Processing Unit, GPU) 1041 and a microphone 1042, the graphics processor 1041 processing image data of still pictures or video obtained by an image capturing device (e.g., a camera) in a video capturing mode or an image capturing mode. The processed image frames may be displayed on the display unit 106. The image frames processed by the graphics processor 1041 may be stored in the memory 109 (or other storage medium) or transmitted via the radio frequency unit 101 or the WiFi module 102. The microphone 1042 can receive sound (audio data) via the microphone 1042 in a phone call mode, a recording mode, a voice recognition mode, and the like, and can process such sound into audio data. The processed audio (voice) data may be converted into a format output that can be transmitted to the mobile communication base station via the radio frequency unit 101 in the case of a telephone call mode. The microphone 1042 may implement various types of noise cancellation (or suppression) algorithms to cancel (or suppress) noise or interference generated in the course of receiving and transmitting the audio signal.
In one embodiment, the wearable device 100 includes one or more cameras, and by opening the cameras, capturing of images, photographing, video recording and other functions can be achieved, and the positions of the cameras can be set as required.
The wearable device 100 further comprises at least one sensor 105, such as a light sensor, a motion sensor, and other sensors. Specifically, the light sensor includes an ambient light sensor and a proximity sensor, wherein the ambient light sensor can adjust the brightness of the display panel 1061 according to the brightness of ambient light, and the proximity sensor can turn off the display panel 1061 and/or the backlight when the wearable device 100 moves to the ear. As one type of motion sensor, the accelerometer sensor can detect the acceleration in all directions (typically three axes), and can detect the gravity and direction when stationary, and can be used for applications for recognizing the gesture of a mobile phone (such as horizontal-vertical screen switching, related games, magnetometer gesture calibration), vibration recognition related functions (such as pedometer, knocking), and the like.
In one embodiment, the wearable device 100 further comprises a proximity sensor, by employing the proximity sensor, the wearable device is able to achieve non-contact manipulation, providing more modes of operation.
In one embodiment, the wearable device 100 further comprises a heart rate sensor, which when worn, enables detection of heart rate by being in close proximity to the user.
In one embodiment, the wearable device 100 may further include a fingerprint sensor, by reading a fingerprint, security verification or the like can be achieved.
The display unit 106 is used to display information input by a user or information provided to the user. The display unit 106 may include a display panel 1061, and the display panel 1061 may be configured in the form of a liquid crystal display (Liquid Crystal Display, LCD), an Organic Light-Emitting Diode (OLED), or the like.
In one embodiment, the display panel 1061 employs a flexible display screen, and the wearable device employing the flexible display screen is capable of bending when worn, thereby fitting more. Optionally, the flexible display screen may be an OLED screen body and a graphene screen body, and in other embodiments, the flexible display screen may also be other display materials, which is not limited to this embodiment.
In one embodiment, the display panel 1061 of the wearable device may take a rectangular shape for ease of wrapping when worn. In other embodiments, other approaches may be taken as well.
The user input unit 107 may be used to receive input numeric or character information and to generate key signal inputs related to user settings and function control of the wearable device. In particular, the user input unit 107 may include a touch panel 1071 and other input devices 1072. The touch panel 1071, also referred to as a touch screen, may collect touch operations thereon or thereabout by a user (e.g., operations of the user on the touch panel 1071 or thereabout by using any suitable object or accessory such as a finger, a stylus, etc.) and drive the corresponding connection device according to a predetermined program. The touch panel 1071 may include two parts of a touch detection device and a touch controller. The touch detection device detects the touch azimuth of a user, detects a signal brought by touch operation and transmits the signal to the touch controller; the touch controller receives touch information from the touch detection device, converts it into touch point coordinates, and sends the touch point coordinates to the processor 110, and can receive and execute commands sent from the processor 110. Further, the touch panel 1071 may be implemented in various types such as resistive, capacitive, infrared, and surface acoustic wave. The user input unit 107 may include other input devices 1072 in addition to the touch panel 1071. In particular, other input devices 1072 may include, but are not limited to, one or more of a physical keyboard, function keys (e.g., volume control keys, switch keys, etc.), a trackball, mouse, joystick, etc., as specifically not limited herein.
In one embodiment, the sides of the wearable device 100 may be provided with one or more buttons. The button can realize a plurality of modes such as short pressing, long pressing, rotation and the like, thereby realizing a plurality of operation effects. The number of the buttons can be multiple, and different buttons can be combined for use, so that multiple operation functions are realized.
Further, the touch panel 1071 may overlay the display panel 1061, and when the touch panel 1071 detects a touch operation thereon or thereabout, the touch panel 1071 is transferred to the processor 110 to determine the type of touch event, and then the processor 110 provides a corresponding visual output on the display panel 1061 according to the type of touch event. Although in fig. 1, the touch panel 1071 and the display panel 1061 are two independent components for implementing the input and output functions of the wearable device, in some embodiments, the touch panel 1071 may be integrated with the display panel 1061 to implement the input and output functions of the wearable device, which is not limited herein. For example, when a message notification of a certain application is received through the rf unit 101, the processor 110 may control the message notification to be displayed in a certain preset area of the display panel 1061, where the preset area corresponds to a certain area of the touch panel 1071, and may control the message notification displayed in the corresponding area on the display panel 1061 by performing a touch operation on the certain area of the touch panel 1071.
The interface unit 108 serves as an interface through which at least one external device can be connected with the wearable apparatus 100. For example, the external devices may include a wired or wireless headset port, an external power (or battery charger) port, a wired or wireless data port, a memory card port, a port for connecting a device having an identification module, an audio input/output (I/O) port, a video I/O port, an earphone port, and the like. The interface unit 108 may be used to receive input (e.g., data information, power, etc.) from an external device and transmit the received input to one or more elements within the wearable apparatus 100 or may be used to transmit data between the wearable apparatus 100 and the external device.
In one embodiment, the interface unit 108 of the wearable device 100 adopts a contact structure, and is connected with other corresponding devices through the contact, so as to realize functions of charging, connection and the like. The contact can also be waterproof.
Memory 109 may be used to store software programs as well as various data. The memory 109 may mainly include a storage program area that may store an operating system, application programs required for at least one function (such as a sound playing function, an image playing function, etc.), and a storage data area; the storage data area may store data (such as audio data, phonebook, etc.) created according to the use of the handset, etc. In addition, memory 109 may include high-speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other volatile solid-state storage device.
The processor 110 is a control center of the wearable device, connects various parts of the entire wearable device with various interfaces and lines, performs various functions of the wearable device and processes data by running or executing software programs and/or modules stored in the memory 109, and invoking data stored in the memory 109, thereby performing overall monitoring of the wearable device. Processor 110 may include one or more processing units; preferably, the processor 110 may integrate an application processor that primarily handles operating systems, user interfaces, applications, etc., with a modem processor that primarily handles wireless communications. It will be appreciated that the modem processor described above may not be integrated into the processor 110.
The wearable device 100 may further include a power source 111 (such as a battery) for powering the various components, and preferably, the power source 111 may be logically connected to the processor 110 through a power management system, so as to perform functions of managing charging, discharging, and power consumption management through the power management system.
Although not shown in fig. 1, the wearable device 100 may further include a bluetooth module or the like, which is not described herein. The wearable device 100 can be connected with other terminal devices through bluetooth to realize communication and information interaction.
Fig. 2 to fig. 4 are schematic structural diagrams of a wearable device according to an embodiment of the present application. The wearable device comprises a flexible screen. When the wearable device is unfolded, the flexible screen is in a strip shape; when the wearable device is in a wearing state, the flexible screen is bent to be annular. Fig. 2 and 3 show schematic structural diagrams of the wearable device screen when unfolded, and fig. 4 shows schematic structural diagrams of the wearable device screen when bent.
Based on the above embodiments, it can be seen that if the device is a wristwatch, a bracelet, or a wearable device, the screen of the device may not cover the watchband area of the device, or may cover the watchband area of the device. The application proposes an alternative embodiment, in which the device may be a wristwatch, a bracelet or a wearable device, comprising a screen and a connection. The screen may be a flexible screen and the connection may be a wristband. Alternatively, the screen of the device or the display area of the screen may be partially or fully overlaid on the wristband of the device. Fig. 5 is a schematic hardware diagram of an implementation manner of a wearable device according to an embodiment of the present application, where a screen of the device extends to two sides, and a part of the screen is covered on a watchband of the device. In other embodiments, the screen of the device may also be entirely covered on the watchband of the device, which is not limited to the embodiment of the present application.
Example 1
Fig. 6 is a flowchart of a first embodiment of a screenshot control method in a video capture process according to the present invention. A screenshot control method in a video shooting process, the method comprising:
s1, acquiring a wearing state and a shooting state of a wearing device, determining a sight range of the wearing device according to the wearing state, and determining a shooting azimuth of the wearing device according to the shooting state;
s2, dividing the video preview area in the display area in the sight range, and dividing the video preview area in the display area outside the sight range to obtain a screenshot preview area;
s3, adjusting the video preview area and the screenshot preview area in real time according to the dynamic change state of the shooting azimuth;
and S4, acquiring a screenshot instruction of a control area of the video area in the shooting process, generating a video screenshot according to the screenshot instruction, and displaying the video screenshot in the screenshot preview area in an arrangement mode.
In this embodiment, first, a wearing state and a shooting state of a wearable device are obtained, a line-of-sight range of the wearable device is determined according to the wearing state, and a shooting azimuth of the wearable device is determined according to the shooting state; then, dividing a video preview area in a display area in the sight line range, and dividing a screenshot preview area in a display area outside the sight line range; then, according to the dynamic change state of the shooting azimuth, the video preview area and the screenshot preview area are adjusted in real time; and finally, in the shooting process, acquiring a screenshot instruction of a control area of the video area, generating a video screenshot according to the screenshot instruction, and displaying the video screenshot in the screenshot preview area in an arrangement mode.
In the prior art, particularly when shooting preview is performed on a wearable device, because the display area of the wearable device is relatively narrow and because the video recording preview area has a specific aspect ratio, even though the video recording preview area is in the display area of the wearable device and has a relatively wide transverse display space, the video recording preview area is relatively narrow due to the deficiency of a longitudinal display space, in the video recording preview process, a user displays and updates screenshot images in the video recording preview area in real time in other side-by-side areas of the video recording preview area, wherein in the video recording process, a pressing instruction in the video recording preview area is acquired, when the pressure sensing value of the pressing instruction is larger than a first threshold value, the current video recording image is intercepted, and meanwhile, the video recording image is transmitted to another display area (namely, a buckling position display area of a wrist strap) which is located in a back direction with the current display position. Therefore, a more convenient video recording interactive operation scheme is realized, so that when a user records videos, the display space of the wearable device is fully utilized, historical screenshot image viewing is conveniently carried out, inconvenience of repeated switching in a narrow display area is avoided, the operation efficiency is improved, and the user experience is enhanced.
The method has the advantages that the wearing state and the shooting state of the wearing equipment are obtained, the sight range of the wearing equipment is determined according to the wearing state, and the shooting direction of the wearing equipment is determined according to the shooting state; then, dividing a video preview area in a display area in the sight line range, and dividing a screenshot preview area in a display area outside the sight line range; then, according to the dynamic change state of the shooting azimuth, the video preview area and the screenshot preview area are adjusted in real time; and finally, in the shooting process, acquiring a screenshot instruction of a control area of the video area, generating a video screenshot according to the screenshot instruction, and displaying the video screenshot in the screenshot preview area in an arrangement mode. The screenshot control scheme in the humanized shooting process is realized, so that when a user records a video through the wearable device, the display space of the wearable device is fully utilized, historical screenshot image viewing is conveniently carried out, inconvenience of repeated switching in a narrow display area is avoided, the operation efficiency is improved, and the user experience is enhanced.
Example two
Fig. 7 is a flowchart of a second embodiment of a screenshot control method in a video capturing process according to the present invention, based on the foregoing embodiment, the obtaining a wearing state and a capturing state of a wearable device, determining a line of sight range of the wearable device according to the wearing state, and determining a capturing azimuth of the wearable device according to the capturing state, including:
s11, acquiring the wearing state according to a pressure sensing component of the wearing equipment, wherein the wearing state comprises a wearing position and an operating area;
s12, determining the shooting state according to the wearing position of the wearing equipment and the shooting component of the wearing equipment, wherein the shooting state comprises shooting directions and the shooting component adopted.
In this embodiment, first, the wearing state is obtained according to a pressure sensing component of the wearing device, where the wearing state includes a wearing position and an operation area; then, determining the shooting state according to the wearing position of the wearing equipment and the shooting component of the wearing equipment, wherein the shooting state comprises a shooting direction and the adopted shooting component.
Optionally, the wearing state is obtained by acquiring pressure sensing information according to a pressure sensing component of the wearing device, which is arranged at the inner side of the device, and calculating the wearing state according to the pressure sensing information, wherein the wearing state comprises a wearing position and an operating area, specifically, the wearing position is obtained by calculating the pressure applied to the pressure sensing component by bones or tissues under the wearing position, and the operating area is obtained by the operating range of the other arm under the wearing position;
optionally, the shooting state is determined according to the wearing position of the wearable device and the shooting component of the wearable device, where the shooting state includes a shooting direction and the shooting component adopted, and specifically, the shooting state is used for determining a scene to which the shooting component of the current wearable device is applied, so that the influence of subsequent screenshot on shooting is avoided.
The method has the beneficial effects that the wearing state is obtained through the pressure sensing component of the wearing equipment, wherein the wearing state comprises a wearing position and an operating area; then, determining the shooting state according to the wearing position of the wearing equipment and the shooting component of the wearing equipment, wherein the shooting state comprises a shooting direction and the adopted shooting component. The screenshot control scheme in the shooting process is more humanized, so that when a user records a video through the wearable device, the display space of the wearable device is fully utilized, historical screenshot image viewing is conveniently carried out, inconvenience of repeated switching in a narrow display area is avoided, the operation efficiency is improved, and the user experience is enhanced.
Example III
Fig. 8 is a flowchart of a third embodiment of a screenshot control method in a video capturing process according to the present invention, based on the above embodiment, the dividing into a video preview area in the display area within the line of sight range, and the dividing into a screenshot preview area in the display area outside the line of sight range includes:
s21, determining the sight line range according to the wearing position;
s22, dividing the display area of the wearable device in the sight line range according to the shooting state to obtain the video preview area.
In this embodiment, first, the line-of-sight range is determined according to the wearing position; then, in the sight range, dividing the display area of the wearable device according to the shooting state to obtain the video preview area.
Optionally, determining the sight line range of the user himself or the recorded object according to the wearing position, wherein in the video recording process, a screenshot instruction of the user himself or the recorded object can be obtained, in the former case, the sight line range of the user himself or the recorded object is determined, and in the latter case, the sight line range of the recorded object is determined;
Optionally, in the line of sight range, according to the shooting state, the display area of the wearable device is divided to obtain the video preview area, and it can be understood that the video preview area is located in the line of sight range of the user or the recorded object, so that the user or the recorded object can conveniently view in real time.
The beneficial effect of this embodiment is that the line of sight range is determined by the wearing position; then, in the sight range, dividing the display area of the wearable device according to the shooting state to obtain the video preview area. The screenshot control scheme in the shooting process is more humanized, so that when a user records a video through the wearable device, the display space of the wearable device is fully utilized, historical screenshot image viewing is conveniently carried out, inconvenience of repeated switching in a narrow display area is avoided, the operation efficiency is improved, and the user experience is enhanced.
Example IV
Fig. 9 is a flowchart of a fourth embodiment of a screenshot control method in a video capturing process according to the present invention, based on the foregoing embodiment, the dividing into a video preview area in the display area within the line of sight range, and the dividing into a screenshot preview area in the display area outside the line of sight range, and further includes:
S23, shielding a touch signal in a display area outside the sight line range;
and S24, dividing the screen capturing preview area in the display range of the shielding touch signal, wherein the initial area of the screen capturing preview area is the central area of the display area outside the sight line range.
In this embodiment, first, in a display area outside the line-of-sight range, a touch signal is shielded; and dividing the screen capturing preview area in the display range of the shielding touch signal, wherein the initial area of the screen capturing preview area is the central area of the display area outside the sight line range.
Optionally, in order to avoid that a display area outside the line of sight is touched by mistake, in this embodiment, touch signals are shielded in the display area outside the line of sight;
optionally, in order to avoid the influence of the screenshot on the video preview area in the view line range, in this embodiment, the screenshot preview area is obtained by dividing in the display range of the shielding touch signal, where the starting area of the screenshot preview area is the center area of the display area outside the view line range.
The embodiment has the beneficial effects that the touch signal is shielded through the display area outside the sight line range; and dividing the screen capturing preview area in the display range of the shielding touch signal, wherein the initial area of the screen capturing preview area is the central area of the display area outside the sight line range. The screenshot control scheme in the shooting process is more humanized, so that when a user records a video through the wearable device, the display space of the wearable device is fully utilized, historical screenshot image viewing is conveniently carried out, inconvenience of repeated switching in a narrow display area is avoided, the operation efficiency is improved, and the user experience is enhanced.
Example five
Fig. 10 is a flowchart of a fifth embodiment of a screenshot control method in a video capturing process according to the present invention, based on the above embodiment, the adjusting the video preview area and the screenshot preview area in real time according to the dynamic change state of the capturing orientation includes:
s31, acquiring a dynamic change state of the shooting azimuth, wherein the dynamic change state comprises a change parameter of a shooting angle and a change parameter of an adopted shooting assembly;
s32, determining a corresponding adjustment reference value according to the dynamic change state, wherein the shooting angle is used as a moving direction reference value of the video preview area and the screenshot preview area.
In this embodiment, first, a dynamic change state of the shooting azimuth is obtained, where the dynamic change state includes a change parameter of a shooting angle and a change parameter of an adopted shooting component; and then, determining a corresponding adjustment reference value according to the dynamic change state, wherein the shooting angle is used as a moving direction reference value of the video preview area and the screenshot preview area.
Optionally, in an actual video recording scene, the wearing device may be in a motion state, specifically, motion parameters such as an angle, a moving speed and the like of the wearing device may change along with time, and meanwhile, when the wearing device has multiple groups of camera assemblies according to different shooting requirements, the currently adopted camera assemblies may also change correspondingly, so in order to enable the video preview area and the screenshot preview area to be in a sight line range and out of the sight line range at the same time, in this embodiment, a dynamic change state of the shooting direction is obtained in real time, where the dynamic change state includes a change parameter of a shooting angle and a change parameter of the adopted shooting assembly;
Optionally, determining a corresponding adjustment reference value according to the dynamic change state, where the shooting angle is used as a moving direction reference value of the video preview area and the screenshot preview area, so that the video preview area and the screenshot preview area in the embodiment are located in a sight line range and outside the sight line range at the same time.
The method has the advantages that the dynamic change state of the shooting azimuth is obtained, wherein the dynamic change state comprises the change parameters of the shooting angle and the change parameters of the adopted shooting assembly; and then, determining a corresponding adjustment reference value according to the dynamic change state, wherein the shooting angle is used as a moving direction reference value of the video preview area and the screenshot preview area. The screenshot control scheme in the shooting process is more humanized, so that when a user records a video through the wearable device, the display space of the wearable device is fully utilized, historical screenshot image viewing is conveniently carried out, inconvenience of repeated switching in a narrow display area is avoided, the operation efficiency is improved, and the user experience is enhanced.
Example six
Fig. 11 is a flowchart of a sixth embodiment of a screenshot control method in a video capturing process according to the present invention, based on the foregoing embodiment, the adjusting the video preview area and the screenshot preview area in real time according to the dynamic change state of the capturing orientation further includes:
s33, respectively moving the video preview area and the screenshot preview area to target positions according to the moving direction reference value;
and S34, respectively taking the corresponding target positions as display centers, and adjusting the region ranges of the video preview region and the screenshot preview region.
In this embodiment, first, according to the moving direction reference value, the video preview area and the screenshot preview area are respectively moved to a target position; and then, respectively taking the corresponding target positions as display centers, and adjusting the region ranges of the video preview region and the screenshot preview region.
Optionally, moving the video preview area and the screenshot preview area to target positions respectively according to the moving direction reference value, wherein the center positions of the video preview area and the screenshot preview area are moved to target positions respectively;
Optionally, the area ranges of the video preview area and the screenshot preview area are adjusted by using the corresponding target positions as display centers, respectively, and it is understood that the area ranges of the video preview area and the screenshot preview area are variable, and particularly, when the screenshot increases, the area range of the screenshot preview area also increases synchronously, but the area still expands on both sides with the display centers.
The method has the advantages that the video preview area and the screenshot preview area are respectively moved to target positions through the moving direction reference value; and then, respectively taking the corresponding target positions as display centers, and adjusting the region ranges of the video preview region and the screenshot preview region. The screenshot control scheme in the shooting process is more humanized, so that when a user records a video through the wearable device, the display space of the wearable device is fully utilized, historical screenshot image viewing is conveniently carried out, inconvenience of repeated switching in a narrow display area is avoided, the operation efficiency is improved, and the user experience is enhanced.
Example seven
Fig. 12 is a flowchart of a seventh embodiment of a screenshot control method in a video shooting process according to the present invention, based on the foregoing embodiment, where in the shooting process, a screenshot instruction of a control area of the video area is obtained, a video screenshot is generated according to the screenshot instruction, and the video screenshot is arranged and displayed in the screenshot preview area, where the method includes:
s41, acquiring a screenshot instruction in the video area in the shooting process, wherein the screenshot instruction comprises a screenshot instruction generated by a touch signal;
s42, determining a target video frame according to the trigger time of the screenshot instruction, and synthesizing multi-frame images in the front-rear range of the target video frame to obtain the video screenshot.
In this embodiment, first, in the shooting process, a screenshot instruction in the video area is obtained, where the screenshot instruction includes a screenshot instruction generated by a touch signal; and then, determining a target video frame according to the trigger time of the screenshot instruction, and synthesizing multi-frame images in the front and rear ranges of the target video frame to obtain the video screenshot.
Optionally, in the shooting process, acquiring a screenshot instruction in the video area, where the screenshot instruction includes a screenshot instruction generated by a touch signal or a long-press touch signal;
Optionally, determining a target video frame according to the triggering moment of the screenshot instruction, when the target video frame is blurred or the edge sharpness of a shooting object is low, acquiring multi-frame images in the front-rear range of the target video frame, and then synthesizing the multi-frame images in the front-rear range of the target video frame to obtain the video screenshot.
The method has the advantages that the screenshot instruction in the video area is obtained in the shooting process, wherein the screenshot instruction comprises a screenshot instruction generated by a touch signal; and then, determining a target video frame according to the trigger time of the screenshot instruction, and synthesizing multi-frame images in the front and rear ranges of the target video frame to obtain the video screenshot. The screenshot control scheme in the shooting process is more humanized, so that when a user records a video through the wearable device, the display space of the wearable device is fully utilized, historical screenshot image viewing is conveniently carried out, inconvenience of repeated switching in a narrow display area is avoided, the operation efficiency is improved, and the user experience is enhanced.
Example eight
Fig. 13 is a flowchart of an eighth embodiment of a screenshot control method in a video shooting process according to the present invention, based on the foregoing embodiment, where in the shooting process, a screenshot instruction of a control area of the video area is obtained, a video screenshot is generated according to the screenshot instruction, and the video screenshot is arranged and displayed in the screenshot preview area, and further includes:
S43, moving the first video screenshot to a display center of the screenshot preview area;
s44, sequentially arranging and displaying the subsequent video screenshots on the two side areas of the display center of the screenshot preview area according to a time axis.
In this embodiment, first, the first video screenshot is moved to the display center of the screenshot preview area; and then, arranging and displaying the subsequent video screenshot on the two side areas of the display center of the screenshot preview area in sequence according to a time axis.
Optionally, moving the first video screenshot to a display center of the screenshot preview area;
optionally, the subsequent video shots are sequentially arranged and displayed on two side areas of the display center of the shot preview area in a time axis, after shooting is completed, the video preview area is released, and meanwhile, the video shots sequentially arranged and displayed on two side areas of the display center of the shot preview area in the time axis are concentrated and overlapped to the original video preview area, so that a video album is generated;
optionally, according to rotation or movement parameters of the wearable device, selecting a corresponding position in the video album and expanding a video screenshot corresponding to the position.
The method has the advantages that the first video screenshot is moved to the display center of the screenshot preview area; and then, arranging and displaying the subsequent video screenshot on the two side areas of the display center of the screenshot preview area in sequence according to a time axis. The screenshot control scheme in the shooting process is more humanized, so that when a user records a video through the wearable device, the display space of the wearable device is fully utilized, historical screenshot image viewing is conveniently carried out, inconvenience of repeated switching in a narrow display area is avoided, the operation efficiency is improved, and the user experience is enhanced.
Example nine
Based on the above embodiment, the present invention further provides a screenshot control device in a video capturing process, where the device includes:
a memory, a processor, and a computer program stored on the memory and executable on the processor;
the computer program implementing the steps of the method according to any of the preceding claims when executed by the processor.
Specifically, in this embodiment, firstly, a wearing state and a shooting state of a wearable device are obtained, a line-of-sight range of the wearable device is determined according to the wearing state, and a shooting azimuth of the wearable device is determined according to the shooting state; then, dividing a video preview area in a display area in the sight line range, and dividing a screenshot preview area in a display area outside the sight line range; then, according to the dynamic change state of the shooting azimuth, the video preview area and the screenshot preview area are adjusted in real time; and finally, in the shooting process, acquiring a screenshot instruction of a control area of the video area, generating a video screenshot according to the screenshot instruction, and displaying the video screenshot in the screenshot preview area in an arrangement mode.
In the prior art, particularly when shooting preview is performed on a wearable device, because the display area of the wearable device is relatively narrow and because the video recording preview area has a specific aspect ratio, even though the video recording preview area is in the display area of the wearable device and has a relatively wide transverse display space, the video recording preview area is relatively narrow due to the deficiency of a longitudinal display space, in the video recording preview process, a user displays and updates screenshot images in the video recording preview area in real time in other side-by-side areas of the video recording preview area, wherein in the video recording process, a pressing instruction in the video recording preview area is acquired, when the pressure sensing value of the pressing instruction is larger than a first threshold value, the current video recording image is intercepted, and meanwhile, the video recording image is transmitted to another display area (namely, a buckling position display area of a wrist strap) which is located in a back direction with the current display position. Therefore, a more convenient video recording interactive operation scheme is realized, so that when a user records videos, the display space of the wearable device is fully utilized, historical screenshot image viewing is conveniently carried out, inconvenience of repeated switching in a narrow display area is avoided, the operation efficiency is improved, and the user experience is enhanced.
The method has the advantages that the wearing state and the shooting state of the wearing equipment are obtained, the sight range of the wearing equipment is determined according to the wearing state, and the shooting direction of the wearing equipment is determined according to the shooting state; then, dividing a video preview area in a display area in the sight line range, and dividing a screenshot preview area in a display area outside the sight line range; then, according to the dynamic change state of the shooting azimuth, the video preview area and the screenshot preview area are adjusted in real time; and finally, in the shooting process, acquiring a screenshot instruction of a control area of the video area, generating a video screenshot according to the screenshot instruction, and displaying the video screenshot in the screenshot preview area in an arrangement mode. The screenshot control scheme in the humanized shooting process is realized, so that when a user records a video through the wearable device, the display space of the wearable device is fully utilized, historical screenshot image viewing is conveniently carried out, inconvenience of repeated switching in a narrow display area is avoided, the operation efficiency is improved, and the user experience is enhanced.
Examples ten
Based on the above embodiments, the present invention also proposes a computer readable storage medium having a bitmap processing program stored thereon, which when executed by a processor implements the steps of the bitmap processing method according to any one of the above.
By means of the bitmap processing method, the bitmap processing device and the computer readable storage medium, the wearing state and the shooting state of the wearing device are obtained, the sight range of the wearing device is determined according to the wearing state, and the shooting direction of the wearing device is determined according to the shooting state; then, dividing a video preview area in a display area in the sight line range, and dividing a screenshot preview area in a display area outside the sight line range; then, according to the dynamic change state of the shooting azimuth, the video preview area and the screenshot preview area are adjusted in real time; and finally, in the shooting process, acquiring a screenshot instruction of a control area of the video area, generating a video screenshot according to the screenshot instruction, and displaying the video screenshot in the screenshot preview area in an arrangement mode. The screenshot control scheme in the humanized shooting process is realized, so that when a user records a video through the wearable device, the display space of the wearable device is fully utilized, historical screenshot image viewing is conveniently carried out, inconvenience of repeated switching in a narrow display area is avoided, the operation efficiency is improved, and the user experience is enhanced.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising one … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element.
The foregoing embodiment numbers of the present invention are merely for the purpose of description, and do not represent the advantages or disadvantages of the embodiments.
From the above description of the embodiments, it will be clear to those skilled in the art that the above-described embodiment method may be implemented by means of software plus a necessary general hardware platform, but of course may also be implemented by means of hardware, but in many cases the former is a preferred embodiment. Based on such understanding, the technical solution of the present invention may be embodied essentially or in a part contributing to the prior art in the form of a software product stored in a storage medium (e.g. ROM/RAM, magnetic disk, optical disk) comprising instructions for causing a terminal (which may be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.) to perform the method according to the embodiments of the present invention.
The embodiments of the present invention have been described above with reference to the accompanying drawings, but the present invention is not limited to the above-described embodiments, which are merely illustrative and not restrictive, and many forms may be made by those having ordinary skill in the art without departing from the spirit of the present invention and the scope of the claims, which are to be protected by the present invention.

Claims (5)

1. A method for controlling screenshot in a video shooting process, the method comprising:
acquiring a wearing state and a shooting state of a wearing device, determining a sight line range of the wearing device according to the wearing state, and determining a shooting azimuth of the wearing device according to the shooting state;
dividing a video preview area in a display area in the sight range, and dividing a screenshot preview area in a display area outside the sight range;
according to the dynamic change state of the shooting azimuth, the video preview area and the screenshot preview area are adjusted in real time;
in the shooting process, acquiring a screenshot instruction of a control area of the video area, generating a video screenshot according to the screenshot instruction, and arranging and displaying the video screenshot in the screenshot preview area;
Dividing the display area within the sight line range to obtain a video preview area, and dividing the display area outside the sight line range to obtain a screenshot preview area, wherein the method comprises the following steps:
shielding a touch signal in a display area outside the sight line range;
dividing the screen capturing preview area in the display range of the shielding touch signal, wherein the initial area of the screen capturing preview area is the central area of the display area outside the sight line range;
the adjusting the video preview area and the screenshot preview area in real time according to the dynamic change state of the shooting azimuth includes:
acquiring a dynamic change state of the shooting azimuth, wherein the dynamic change state comprises a change parameter of a shooting angle and a change parameter of an adopted shooting assembly;
determining a corresponding adjustment reference value according to the dynamic change state, wherein the shooting angle is used as a moving direction reference value of the video preview area and the screenshot preview area;
respectively moving the video preview area and the screenshot preview area to target positions according to the moving direction reference value;
Respectively taking the corresponding target positions as display centers, and adjusting the region ranges of the video preview region and the screenshot preview region;
when the screen shots are increased, the area range of the screen shot preview area is synchronously increased, and the screen shot preview area is expanded at two sides by the display center;
in the shooting process, acquiring a screenshot instruction of a control area of the video area, generating a video screenshot according to the screenshot instruction, and displaying the video screenshot in the screenshot preview area in a distributed manner, wherein the method comprises the following steps:
acquiring a screenshot instruction in the video area in the shooting process, wherein the screenshot instruction comprises a screenshot instruction generated by a touch signal;
determining a target video frame according to the triggering moment of the screenshot instruction, and synthesizing multi-frame images in the front and rear ranges of the target video frame to obtain the video screenshot;
in the shooting process, acquiring a screenshot instruction of a control area of the video area, generating a video screenshot according to the screenshot instruction, and displaying the video screenshot in the screenshot preview area in a distributed manner, wherein the method comprises the following steps:
moving the first video screenshot to a display center of the screenshot preview area;
Sequentially arranging and displaying the subsequent video screenshots on the two side areas of the display center of the screenshot preview area according to a time axis;
wherein,,
in the video recording preview process, the user displays and updates the video screenshot in the recording process in real time in other side-by-side areas of the video recording preview area;
transmitting the video screenshot to another display area which is opposite to the current display direction, wherein the other display area is a display area at the buckling part of the wrist strap of the wearable device;
when a plurality of video shots are sequentially intercepted, the video shots are respectively sent to the other back display area, the video shots in the other back display area are sequentially moved to the forward display area from the other back display area, and when the video recording is completed, the historical video shots are checked according to the time axis of the video shots.
2. The screenshot control method in a video capturing process according to claim 1, wherein the acquiring the wearing state and the capturing state of the wearable device, determining the line of sight range of the wearable device according to the wearing state, and determining the capturing azimuth of the wearable device according to the capturing state includes:
Acquiring the wearing state according to a pressure sensing component of the wearing device, wherein the wearing state comprises a wearing position and an operating area;
and determining the shooting state according to the wearing position of the wearing equipment and the shooting component of the wearing equipment, wherein the shooting state comprises a shooting azimuth and the shooting component adopted.
3. The method according to claim 2, wherein the dividing the display area within the line of sight to obtain the video preview area and the dividing the display area outside the line of sight to obtain the screenshot preview area includes:
determining the line of sight according to the wearing position;
and dividing the display area of the wearable equipment according to the shooting state in the sight range to obtain the video preview area.
4. A screenshot control apparatus in a video capturing process, the apparatus comprising:
a memory, a processor, and a computer program stored on the memory and executable on the processor;
the computer program implementing the steps of the method according to any one of claims 1 to 3 when executed by the processor.
5. A computer-readable storage medium, wherein a screenshot control program in a video capturing process is stored on the computer-readable storage medium, which when executed by a processor, implements the steps of the screenshot control method in a video capturing process as claimed in any one of claims 1 to 3.
CN201910917201.0A 2019-09-26 2019-09-26 Screenshot control method, equipment and computer readable storage medium in video shooting process Active CN110620875B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910917201.0A CN110620875B (en) 2019-09-26 2019-09-26 Screenshot control method, equipment and computer readable storage medium in video shooting process

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910917201.0A CN110620875B (en) 2019-09-26 2019-09-26 Screenshot control method, equipment and computer readable storage medium in video shooting process

Publications (2)

Publication Number Publication Date
CN110620875A CN110620875A (en) 2019-12-27
CN110620875B true CN110620875B (en) 2023-08-11

Family

ID=68924274

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910917201.0A Active CN110620875B (en) 2019-09-26 2019-09-26 Screenshot control method, equipment and computer readable storage medium in video shooting process

Country Status (1)

Country Link
CN (1) CN110620875B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113038218B (en) * 2021-03-19 2022-06-10 厦门理工学院 Video screenshot method, device, equipment and readable storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105898220A (en) * 2016-04-25 2016-08-24 乐视控股(北京)有限公司 Terminal, driving recording system and method for terminal
CN108259973A (en) * 2017-12-20 2018-07-06 青岛海信电器股份有限公司 The display methods of the graphic user interface of smart television and television image sectional drawing
CN110099218A (en) * 2019-05-31 2019-08-06 努比亚技术有限公司 Interaction control method, equipment and computer readable storage medium in a kind of shooting process
CN110177208A (en) * 2019-05-31 2019-08-27 努比亚技术有限公司 A kind of association control method of video record, equipment and computer readable storage medium
CN110177209A (en) * 2019-05-31 2019-08-27 努比亚技术有限公司 A kind of video parameter regulation method, equipment and computer readable storage medium

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP6899510B2 (en) * 2016-04-14 2021-07-07 Line株式会社 Display control method, terminal, and program

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105898220A (en) * 2016-04-25 2016-08-24 乐视控股(北京)有限公司 Terminal, driving recording system and method for terminal
CN108259973A (en) * 2017-12-20 2018-07-06 青岛海信电器股份有限公司 The display methods of the graphic user interface of smart television and television image sectional drawing
CN110099218A (en) * 2019-05-31 2019-08-06 努比亚技术有限公司 Interaction control method, equipment and computer readable storage medium in a kind of shooting process
CN110177208A (en) * 2019-05-31 2019-08-27 努比亚技术有限公司 A kind of association control method of video record, equipment and computer readable storage medium
CN110177209A (en) * 2019-05-31 2019-08-27 努比亚技术有限公司 A kind of video parameter regulation method, equipment and computer readable storage medium

Also Published As

Publication number Publication date
CN110620875A (en) 2019-12-27

Similar Documents

Publication Publication Date Title
CN110299100B (en) Display direction adjustment method, wearable device and computer readable storage medium
CN109639987B (en) Bracelet shooting method and device and computer readable storage medium
CN110099218B (en) Interactive control method and device in shooting process and computer readable storage medium
CN109947249B (en) Interaction method of wearable device, wearable device and computer storage medium
CN110225282B (en) Video recording control method, device and computer readable storage medium
CN110177209B (en) Video parameter regulation and control method, device and computer readable storage medium
CN110187769B (en) Preview image viewing method, equipment and computer readable storage medium
CN109947524B (en) Interface display method, wearable device and computer readable storage medium
CN110933293A (en) Shooting method, terminal and computer readable storage medium
CN110198411B (en) Depth of field control method and device in video shooting process and computer readable storage medium
CN110113529B (en) Shooting parameter regulation and control method and device and computer readable storage medium
CN110177208B (en) Video recording association control method, equipment and computer readable storage medium
CN110072071B (en) Video recording interaction control method, equipment and computer readable storage medium
CN110620875B (en) Screenshot control method, equipment and computer readable storage medium in video shooting process
CN110650289B (en) Shooting depth of field control method, equipment and computer readable storage medium
CN110062167B (en) Linkage shooting control method and device and computer readable storage medium
CN110620876B (en) Image preview interaction method, device and computer readable storage medium
CN110109605B (en) Image sharing method and device and computer readable storage medium
CN110191282B (en) Shooting parameter regulation and control method and device and computer readable storage medium
CN110187950B (en) Method for adjusting picture display position, wearable device and storage medium
CN110049248B (en) Shot object regulation and control method and device and computer readable storage medium
CN110083427B (en) Application program interaction control method, device and computer readable storage medium
CN110049253B (en) Focusing control method and device and computer readable storage medium
CN110083513B (en) Lamp effect control method, equipment and computer readable storage medium
CN110069102B (en) Display area regulation and control method and device and computer readable storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
TA01 Transfer of patent application right

Effective date of registration: 20220530

Address after: 1801, building 2, Chongwen Park, Nanshan Zhiyuan, No. 3370, Liuxian Avenue, Fuguang community, Taoyuan Street, Nanshan District, Shenzhen, Guangdong 518000

Applicant after: NUBIA TECHNOLOGY Co.,Ltd.

Address before: 518000 6-8 / F, 10-11 / F, 6 / F, 6-10 / F, C zone, Han's innovation building, No. 9018, Beihuan Avenue, high tech Zone, Nanshan District, Shenzhen City, Guangdong Province

Applicant before: NUBIA TECHNOLOGY Co.,Ltd.

Applicant before: NANCHANG NUBIA TECHNOLOGY Co.,Ltd.

TA01 Transfer of patent application right
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant