WO2015043403A1 - Method, apparatus, and terminal for obtaining video data - Google Patents

Method, apparatus, and terminal for obtaining video data Download PDF

Info

Publication number
WO2015043403A1
WO2015043403A1 PCT/CN2014/086719 CN2014086719W WO2015043403A1 WO 2015043403 A1 WO2015043403 A1 WO 2015043403A1 CN 2014086719 W CN2014086719 W CN 2014086719W WO 2015043403 A1 WO2015043403 A1 WO 2015043403A1
Authority
WO
WIPO (PCT)
Prior art keywords
video
video data
video segments
segments
time length
Prior art date
Application number
PCT/CN2014/086719
Other languages
French (fr)
Inventor
Yi SHAN
Hui Tang
Original Assignee
Tencent Technology (Shenzhen) Company Limited
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tencent Technology (Shenzhen) Company Limited filed Critical Tencent Technology (Shenzhen) Company Limited
Publication of WO2015043403A1 publication Critical patent/WO2015043403A1/en

Links

Images

Classifications

    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/02Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
    • G11B27/031Electronic editing of digitised analogue information signals, e.g. audio or video signals

Definitions

  • the present application relates to the field of multimedia technologies, and in particular, to a method, an apparatus, and a terminal for obtaining video data.
  • a method for obtaining video data includes triggering a camera apparatus on a terminal to successively capture one piece of video data by clicking a related interface interaction element on an application interface, for example, a button or a touch control option for capturing a video; and using the captured consecutive video data as the obtained video data.
  • the obtained video data is captured by the camera apparatus at one time, and thus, the video capturing manner is monotonous.
  • data that is not required by a user is often captured.
  • the obtained video data includes redundant video data, and the redundancy of the obtained video data is increased.
  • the present application is implemented in a system of obtaining and sharing video data using a social networking application that has one or more processors, memory and one or more modules, programs or sets of instructions stored in the memory for performing multiple functions. Instructions for performing these functions may be included in a computer program product configured for execution by one or more processors.
  • One aspect of the present application involves a computer-implemented method of obtaining and sharing video data using a social networking application performed by a computer having one or more processors and memory.
  • the computer-implemented method includes: recording a plurality of video segments successively; marking each video segment with a respective identifier; caching the plurality of video segments in the memory for processing; determining whether a total time length of the plurality of video segments exceeds a pre-set threshold; stopping video recording if the total time length of the plurality of video segments exceeds the pre-set threshold; and integrating the plurality of video segments into one video according to a predefined order.
  • determining whether the total time length of the plurality of video segments exceeds the pre-set threshold further comprises configuring a timer with a value equal to the pre-set threshold; subtracting time length of the video segment from the value of the timer once the video segment is recorded; determining whether the value of the timer equals to or less than zero; and when the value of the timer equals to or less than zero, instructing that the total time length of the plurality of video segments exceeds the pre-set threshold.
  • integrating the plurality of video segments into the one video according to a predefined order further comprises reading the plurality of identifiers that correspond to the plurality of video segments, respectively; re-organizing the plurality of video segments into a sequence based on sorting of the plurality of identifiers; and aggregating the re-organized plurality of video segments into the one video.
  • integrating the plurality of video segments into the one video according to a predefined order further comprises reading a plurality of recording times that correspond to the plurality of video segments, respectively; re-organizing the plurality of video segments into a sequence based on sorting of the plurality of recording times; and aggregating the re-organized plurality of video segments into the one video.
  • the method of obtaining and sharing video data using a social networking application further comprises prompting a message to a user of the social networking application whether to share the one video in a social network; and if receiving an instruction from the user to share the one video in the social network, choosing a participant of the social network that the user wants to share with, and sending the one video to the participant via the social network application.
  • the system includes memory, one or more processors, and one or more programs stored in the memory and configured for execution by the one or more processors.
  • the one or more programs include: recording a plurality of video segments successively; marking each video segment with a respective identifier; caching the plurality of video segments in the memory for processing; determining whether a total time length of the plurality of video segments exceeds a pre-set threshold; stopping video recording if the total time length of the plurality of video segments exceeds the pre-set threshold; and integrating the plurality of video segments into one video according to a predefined order.
  • Another aspect of the present application involves a non-transitory computer readable storage medium having stored therein one or more programs for execution by one or more processors of a system of obtaining and sharing video data using a social networking application, the one or more programs including instructions for: recording a plurality of video segments successively; marking each video segment with a respective identifier; caching the plurality of video segments in the memory for processing; determining whether a total time length of the plurality of video segments exceeds a pre-set threshold; stopping video recording if the total time length of the plurality of video segments exceeds the pre-set threshold; and integrating the plurality of video segments into one video according to a predefined order.
  • FIG. 1 is a flowchart of a method for obtaining video data according to Embodiment 1 of the present application
  • FIG. 2 is a flowchart of a method for obtaining video data according to Embodiment 2 of the present application
  • FIG. 3 is a first sequence diagram of identifiers corresponding to video data according to Embodiment 2 of the present application;
  • FIG. 4 is a second sequence diagram of identifiers corresponding to video data according to Embodiment 2 of the present application.
  • FIG. 5 is a first schematic structural diagram of an apparatus for obtaining video data according to Embodiment 3 of the present application.
  • FIG. 6 is a second schematic structural diagram of the apparatus for obtaining video data according to Embodiment 3 of the present application.
  • FIG. 7 is a third schematic structural diagram of the apparatus for obtaining video data according to Embodiment 3 of the present application.
  • FIG. 8 is a schematic structural diagram of a terminal according to Embodiment 4 of the present application.
  • the embodiment of the present application provides a method for obtaining video data.
  • the method is applicable to a social network application and installed on a terminal with a capturing function.
  • the terminal includes, but is not limited to, a mobile phone, a computer, a tablet, and the like.
  • a specific form of the terminal is not limited in this embodiment. Referring to FIG. 1, the method provided in this embodiment includes:
  • a t step 101 a plurality of video segments is recorded successively, each video segment is marked with a respective identifier, and then it is determined whether a total time length of the plurality of video segments exceeds a pre-set threshold.
  • the following steps may be performed:
  • the video recording is stopped if the total time length of the plurality of video segments exceeds the pre-set threshold.
  • the plurality of video segments are integrated into one video according to a predefined order.
  • the step of integrating the plurality of video segments into one video according to a predefined order includes:
  • the step of integrating the plurality of video segments into one video according to a predefined order includes:
  • the following steps may be performed:
  • video data is successively captured for multiple time periods, and the video data obtained by successively capturing for the multiple time periods is combined into one piece of video data.
  • one piece of video data is obtained by selectively capturing for multiple time periods, and content of captured video data can be selected at will. Therefore, the manner of obtaining video data is extended, and redundancy of obtained video data is decreased, thereby improving user experience.
  • the embodiment of the present application provides a method for obtaining video data.
  • the method is applicable to a social network application and installed in a terminal with a capturing function.
  • the terminal includes, but is not limited to, a mobile phone, a computer, a tablet, and the like.
  • a specific form of the terminal is not limited in this embodiment.
  • a procedure of the method provided in this embodiment includes:
  • step 201 successively capturing video data for multiple time periods, and determining, each time period for which video data is captured, whether a total time length of the video data obtained by successively capturing for the multiple times reaches a preset time length.
  • the preset time length of capturing video data may be set in advance, for example, by a provider that provides a video data obtaining service.
  • a corresponding operation interface may be provided in the social network application, so that a user of the social network application can set the preset time length by using the operation interface.
  • the time length set by the user is used as a determined preset time length of capturing video data.
  • the preset time length of capturing video data may also be set by selecting other methods, which is not specifically limited in this embodiment.
  • the determined preset time length is also not limited in this embodiment, and includes, but is not limited to, 6 seconds, 8 seconds, or the like.
  • a timer may be set according to the determined preset time length.
  • An initial value of the timer may be the determined preset time length.
  • the step of determining, for each time capturing video data, whether a total time length of the video data obtained by successively capturing for the multiple times reaches a preset time length includes:
  • the initial value of the timer may be set to 6 sec.
  • capturing video data is successively performed multiple times. If a time length of capturing video data for the first time is 2.5 sec, the value of the timer is subtracted by 2.5 sec. At this point, the value of the timer is 3.5. If a time length of capturing video data for the second time is 2 sec, the value of the timer is subtracted by 2 sec. At this time, the value of the timer is 1.5 sec, which means that a remained capturing time length is 1.5 sec. When video data is captured again and a time length reaches 1.5 sec, the value of the timer is subtracted by 1.5 sec.
  • the value of the timer is 0, i. e. , the timer is cleared. In this case, it may be determined that a total time length of video data obtained by successively capturing for three times reaches the preset time length. Therefore, capturing of video data is stopped.
  • the preset time length may be set to a larger value, so that capturing can be successively performed for multiple times to obtain complete video data expected by a user, and capturing is not stopped in the capturing process when the preset time length is reached.
  • multiple manners are available for triggering capturing of video data each time, and may include, but is not limited to, detecting an operation of pressing a capturing button. If it is detected that the capturing button is pressed, a signal of capturing video data is obtained, and capturing of video data is performed accordingly.
  • multiple manners are available for stopping capturing of video data each time, and may include, but is not limited to, detecting an operation of pressing a capturing stopping button.
  • the capturing button and the capturing stopping button may be two buttons set on the terminal.
  • step 202 if the total time length of the video data obtained by successively capturing for the multiple times reaches the preset time length, capturing of video data is stopped.
  • the application corresponding to the social network may stop capturing of video data.
  • the application corresponding to the social network may display related prompt information to notify the user that capturing of video data ends.
  • a dialog box may be displayed on the interface of the social network application, and information displayed in the dialog box is used to notify the user that capturing of video data ends.
  • Other manners may also be selected to prompt that capturing of video data ends. The specific prompt manner is not limited in this embodiment.
  • corresponding prompt information may further be displayed to prompt the user that the capturing process will end soon.
  • a manner of displaying the prompt information and content of the displayed prompt information are also not specifically limited in this embodiment.
  • the method may further provide a related operation interface through which the user can trigger an operation to stop capturing of video data.
  • a manner of providing the related operation interface by the social network application is not specifically limited in this embodiment, and includes, but is not limited to, displaying a corresponding operation option on an interaction interface provided by the application corresponding to the social network, and determining, by detecting a selection operation of the user on the operation option, whether to stop capturing of video data.
  • the interaction interface is a channel for information exchange between the user and the terminal (a computer, a notebook, a mobile terminal, a mobile phone, or the like) .
  • the user uses the interaction interface to input information on the terminal and perform an operation.
  • the terminal uses the interaction interface to provide information to the user for reading, analyzing, and determining by the user.
  • the interaction interface in this embodiment includes a home page displayed when software or an application is opened, and interaction interfaces entered after various function buttons and entries are clicked. Interaction interfaces that appear at different levels or due to different triggering are different. Interaction interfaces entered due to a same triggering or from a same entry are the same regardless of whether reading content on the interfaces changes. Different interfaces generally mean that interaction interfaces at a certain level and displayed to a user are different in the interaction interfaces themselves, and do not mean that the interaction interfaces are different in content.
  • a "combine” option is displayed on an interaction interface provided by the application corresponding to the social network.
  • an indication of stopping capturing video data is obtained, and the two pieces of video data obtained by capturing before is used as video data obtained by successively capturing for multiple times, and a combining operation is performed in a subsequent step.
  • the video data obtained by successively capturing for the multiple times is combined into one piece of video data, and use the video data obtained after combination as obtained video data.
  • the step of combining the video data obtained by successively capturing for the multiple times into one piece of video data includes, but is not limited to, the following two manners:
  • a first combining technique may include combining, according to a time sequence of capturing video data, the video data obtained by successively capturing for the multiple times into one piece of video data.
  • the first combining technique multiple different pieces of video data are obtained after successively capturing for multiple times. Because different pieces of video data are specifically captured at different times in a capturing process, the multiple pieces of video data obtained after successively capturing for the multiple times are sorted according to the specific capturing times. Then, the sorted multiple pieces of video data are combined into one piece of video data, and the video data obtained after combination is used as obtained video data.
  • the manner of sorting, according to the specific capturing times, the multiple pieces of video data obtained after successively capturing for the multiple times is not specifically limited in this embodiment. For example, the multiple pieces of video data may be sorted according to the capturing times in an early-to-late sequence, or in a late-to-early sequence.
  • the multiple pieces of video data when the multiple pieces of video data are combined into one piece of video data, the multiple pieces of video data may be combined according to specific file formats of the video data.
  • the method of combining the video data obtained by successively capturing for the multiple times into one piece of video data is not specifically limited in this embodiment.
  • three pieces of video data are obtained after successively capturing for three times.
  • a first piece is video data V_1 of 2.5 sec
  • a second piece is video data V_2 of 2 sec
  • a third piece is video data V_3 of 1.5 sec.
  • Specific capturing times of the three pieces of video data are respectively 10: 31 am, 10: 27 am, and 10: 40 am in a same day.
  • the three pieces of video data are sorted according to the specific capturing times in an early-to-late sequence, and a final sorted sequence is V_2, V_1, and V_3.
  • the three pieces of video data may be combined in the foregoing sequence, and video data obtained after combination is used as obtained video data.
  • a second combining technique may include displaying identifiers corresponding to the video data obtained by successively capturing for the multiple times, and combining the video data obtained by successively capturing for the multiple times into one piece of video data according to how the user sorts the identifiers corresponding to the video data obtained by successively capturing for the multiple times.
  • an identifier may be set for the piece of video data. Different pieces of video data correspond to different identifiers. After video data is successively captured for multiple times, identifiers corresponding to video data obtained by successively capturing for the multiple times may be displayed on an interface.
  • a social network application may provide a related operation interface. A user may sort, by using the related operation interface, the identifiers corresponding to the video data obtained by successively capturing for the multiple times. Finally, the social network application combines, according to how the user sorts the identifiers, the video data obtained by successively capturing for the multiple times into one piece of video data, and uses video data obtained after combination as obtained video data.
  • a user obtains three pieces of video data after successively capturing video three times with a touch screen mobile phone.
  • a first piece is video data V_1 of 2.5 sec
  • a second piece is video data V_2 of 2 sec
  • a third piece is video data V_3 of 1.5 sec.
  • Identifiers of the three pieces of video data are respectively V_1, V_2, and V_3.
  • the identifiers on an interface and corresponding to the video data obtained by successively capturing video three times are shown in FIG. 3.
  • the user may sort icons corresponding to the three identifiers displayed on the interface shown in FIG. 3, so that the identifiers corresponding to the video data obtained by successively capturing for three times are sorted.
  • a final sorting result is shown in FIG. 4. It can be known from FIG.
  • a first application may combine, according to the foregoing sequence, the video data, V_3, V_2, and V_1, obtained by successively capturing for three times into one piece of video data.
  • the social network application may issue prompt information to prompt a user whether to share the combined video data, for example, a dialog box or a pop-up window may be used, which is not specifically limited in this embodiment.
  • the user may select according to the prompt information whether to share the combined video data.
  • the indication about sharing the combined video data is obtained by detecting a selection result of the user, and a sharing operation is performed according to the indication. A case in which the method provided in this embodiment is applied to a social network application is still used as an example.
  • the social network application may transmit the combined video data to a corresponding terminal according to sharing destination information included in the indication.
  • Specific content of the sharing destination information included in the indication is not limited in this embodiment, and includes, but is not limited to, a user nickname, an account number, and the like registered by the user in the application corresponding to the social network.
  • Objects that shares others' video data and whose video data is shared by others may be users registered in applications corresponding to a same social network, and may also be users registered in applications corresponding to different social networks. That is, the combined video data obtained may be shared on an application platform corresponding to a same social network, or on application platforms corresponding to different social networks.
  • a specific sharing manner and scenario are not limited in this embodiment.
  • a terminal A is associated with an account number A1 registered by a user A in an application corresponding to a first social network
  • a terminal B is associated with an account number B registered by a user B in the application corresponding to the first social network
  • the terminal A is further associated with an account number A2 registered by the user A in an application corresponding to a second social network
  • a terminal C is associated with an account number C registered by a user C in the application corresponding to the second social network
  • an identifier of user A in the application corresponding to the first social network is A1
  • an identifier of user B in the application corresponding to the first social network is B
  • an identifier of user A in the application corresponding to the second social network is A2
  • an identifier of user C in the application corresponding to the second social network is C.
  • terminal A After obtaining the combined video data in the application corresponding to the first social network; terminal A detects that user A selects the identifier B corresponding to user B on an interaction interface provided by the application corresponding to the first social network; and detects that user A selects a sharing option by using an interaction interface provided by the application corresponding to the social network, terminal A then determines that the combined video data needs to be shared with a terminal corresponding to identifier B. Therefore, terminal A sends the combined video data and identifier B to a server of the application corresponding to the first social network. After finding that the terminal corresponding to identifier B is terminal B, the server forwards, to terminal B, the combined video data sent by terminal A. In this way, the combined video data obtained by terminal A is shared with user B corresponding to terminal B.
  • terminal A after terminal A obtains the combined video data in the application corresponding to the first social network, and detects that user A selects the sharing option on the interaction interface provided by the application corresponding to the first social network. Terminal A then displays, on the interaction interface provided by the application corresponding to the social network, applications with which the video data can be shared.
  • the applications with which the video data can be shared include the application corresponding to the second social network.
  • terminal A displays a friend list of user A in the application corresponding to the second social network. User C is displayed in the friend list.
  • terminal A After detecting that user A selects identifier C corresponding to user C, terminal A sends the video data obtained after combination and identifier C to a server of the application corresponding to the second social network. After finding that a terminal corresponding to identifier C is terminal C, the server forwards, to terminal C, the combined video data sent by terminal A. In this way, the video data obtained by terminal A is shared with user C corresponding to terminal C.
  • video data is successively captured multiple times, and the captured video data is combined into one piece of video data.
  • one piece of video data is obtained by selectively capturing video multiple times, and the content of captured video data can be selected at will. Therefore, the manner of obtaining video data is extended, and redundancy of obtained video data is decreased, thereby improving user experience.
  • an embodiment of the present application provides an apparatus for obtaining video data.
  • the apparatus includes:
  • a capturing module 501 configured to successively capture video data multiple times
  • a determining module 502 configured to determine, each time video data is captured, whether a total time length of the video data obtained by successively capturing for the multiple times reaches a preset time length;
  • control module 503 configured to control the capturing module to stop capturing of video data if the total time length of the video data obtained by successively capturing for the multiple times reaches the preset time length;
  • a combining module 504 configured to combine the video data obtained by successively capturing for the multiple times into one piece of video data, and use video data obtained after combination as obtained video data.
  • the apparatus for obtaining video data further includes:
  • a setting module 505 configured to set the preset time length on a timer, where a value of the timer decreases progressively according to a time length of capturing video data, and is cleared when the total time length of capturing video data reaches the preset time length;
  • the determining module 502 is configured to determine, each time video data is captured, whether the value of the timer is cleared; and when the value of the timer is cleared, determine that the total time length of video data obtained by successively capturing for the multiple times reaches the preset time length.
  • the combining module 504 is configured to display identifiers corresponding to the video data obtained by successively capturing for the multiple times, and combine the video data into one piece of video data according to how a user sorts the identifiers corresponding to the video data obtained by successively capturing for the multiple times.
  • the apparatus for obtaining video data further includes:
  • a prompt module 506, configured to prompt whether to share the video data obtained after combination
  • a transmitting module 507 configured to transmit the video data obtained after combination to another terminal after an indication about sharing the combined video data is received.
  • the apparatus provided in this embodiment successively captures video data for multiple times, and combines the video data obtained by successively capturing for the multiple times into one piece of video data.
  • one piece of video data is obtained by selectively capturing for multiple times, and content of captured video data can be selected at will. Therefore, the manner of obtaining video data is extended, and redundancy of obtained video data is decreased, thereby improving user experience.
  • This embodiment provides a terminal.
  • the terminal may be configured to execute the method for obtaining video data provided in the foregoing embodiments.
  • the terminal 600 may include components such as a radio frequency (RF) circuit 110, a memory 120 including one or more computer readable storage media, an input unit 130, a display unit 140, a sensor 150, an audio circuit 160, a wireless fidelity (WiFi) module 170, a processor 180 including one or more processing cores, and a power source 190.
  • RF radio frequency
  • a memory 120 including one or more computer readable storage media
  • an input unit 130 a display unit 140
  • a sensor 150 an audio circuit 160
  • WiFi wireless fidelity
  • processor 180 including one or more processing cores
  • a power source 190 a power source 190.
  • the terminal may include different components, additional components, or fewer components, or any combination of the foregoing than shown in FIG. 8.
  • the RF circuit 110 may be configured to receive and send a signal during communication, or information receiving and sending. In particular, after receiving downlink information of a base station, the RF circuit 110 sends the downlink information to one or more processors 180 for processing. In addition, the RF circuit 110 sends uplink data to the base station.
  • the RF circuit 110 includes, but is not limited to, an antenna, at least one amplifier, a tuner, one or more oscillator, a subscriber identity module (SIM) card, a transceiver, a coupler, a low noise amplifier (LNA) , a duplexer, and the like.
  • SIM subscriber identity module
  • the RF circuit 110 may further communicate with other devices through a network by means of wireless communication.
  • GSM Global System for Mobile Communications
  • GPRS General Packet Radio Service
  • CDMA Code Division Multiple Access
  • WCDMA Wideband Code Division Multiple Access
  • LTE Long Term Evolution
  • SMS Short Messaging Service
  • the memory 120 may be configured to store a software program and module.
  • the processor 180 executes function application and data processing by operating the software program and module stored in the memory 120.
  • the memory 120 may mainly include a program storing area and a data storing area.
  • the program storing area may store an operating system, at least one application required by functions (for example, an audio playing function and an image playing function) , and the like.
  • the data storing area may store data (for example, audio data and a phonebook) that is created according to the use of the terminal 600 and the like.
  • the memory 120 may also include a high-speed random access memory, and may further include a non-volatile memory, for example, include at least one disk storage device, a flash memory, or other volatile solid-state storage devices.
  • the memory 120 may further include a memory controller to control access of the processor 180 and the input unit 130 to the memory 120.
  • the input unit 130 may be configured to receive inputted numerical or character information and generate keyboard, mouse, joystick, optical, and/or trackball signal input related to user setting and function control.
  • the input unit 130 may include a touch-sensitive surface 131 and a secondary input device 132.
  • the touch-sensitive surface 131 is also referred to as a touch display screen or a touch pad.
  • the touch-sensitive surface 131 may collect a touch operation performed by a user on or near the touch-sensitive surface 131 (for example, an operation performed by a user with any proper object or attachment, such as a finger or a touch pen, on or near the touch-sensitive surface 131) , and drive a corresponding connecting apparatus according to a preset program.
  • the touch-sensitive surface 131 may include two parts, a touch detecting apparatus and a touch controller.
  • the touch detecting apparatus detects a touch position of a user and a signal brought by a touch operation, and transmits the signal to the touch controller.
  • the touch controller receives touch information from the touch detecting apparatus, converts the touch information to a touch point coordinate, and sends the touch point coordinate to the processor 180.
  • the touch controller can receive and execute a command sent by the processor 180.
  • the touch-sensitive surface 131 may be implemented in a resistive, capacitive, infrared, or a surface acoustic form, for example.
  • the input unit 130 may further include a secondary input device 132.
  • the secondary input device 132 may include, but is not limited to, one or more of a physical keyboard, a functional key (for example, a volume control key or a switch key) , a trackball, a mouse, a joystick and the like.
  • a functional key for example, a volume control key or a switch key
  • a trackball for example, a mouse, a joystick and the like.
  • the display unit 140 may be configured to display information input by a user, information provided to the user, or various graphic user interfaces of the terminal 600.
  • the graphic user interfaces may include a graph, a text, an icon, a video, and any combination thereof.
  • the display unit 140 may include a display panel 141.
  • the display panel 141 may be configured in a form of a liquid crystal display (LCD) , an organic light-emitting diode (OLED) , or the like.
  • the touch-sensitive surface 131 may cover the display panel 141. After detecting a touch operation performed on or near the touch-sensitive surface 131, the touch-sensitive surface 131 transmits the touch operation to the processor 180 to determine a type of a touch event.
  • the processor 180 provides corresponding visual output on the display panel 141 according to the type of the touch event.
  • the touch-sensitive surface 131 and the display panel 141 implement input and output functions as two independent components, in some embodiments, the touch-sensitive surface 131 and the display panel 141 may be integrated to implement input and output functions.
  • the terminal 600 may further include at least one sensor 150, for example, a light sensor, a motion sensor, and other sensors.
  • the light sensor may include an ambient light sensor and a proximity sensor.
  • the ambient light sensor may adjust brightness of the display panel 141 according to brightness of ambient light.
  • the proximity sensor may turn off the display panel 141 and/or backlight when the terminal 600 moves near an ear of a user.
  • a gravity acceleration sensor may detect magnitudes of accelerations of all directions (generally, three axes) , and can detect a magnitude and direction of gravity in a stationary state.
  • the gravity acceleration sensor may be configured for applications of recognizing a mobile phone posture (for example, switch between portrait and landscape screens, related games, magnetometer calibration) , related functions of vibration recognition (for example, a pedometer and knocking) , and the like.
  • the terminal 600 may further be configured with other sensors, such as a gyroscope, a barometer, a hygrometer, a thermometer, and an infrared sensor, which are not described in detail herein.
  • the audio circuit 160, a loudspeaker 161, a microphone 162 may provide an audio interface between a user and the terminal 600.
  • the audio circuit 160 may transmit an electrical signal obtained by converting received audio data to the loudspeaker 161, so that the loudspeaker 161 converts the electrical signal to a sound signal for output.
  • the microphone 162 converts a collected sound signal to an electrical signal.
  • the audio circuit 160 receives the electrical signal, converts the electrical signal to audio data, and outputs the audio data to the processor 180 for processing. After processing, the audio data is sent through the RF circuit 110 to, for example, another terminal, or is output to the memory 120 for further processing.
  • the audio circuit 160 may further include an earphone jack to provide communication between an external earphone and the terminal 600.
  • WiFi is a short-distance wireless transmission technology.
  • the WiFi module 170 enables a user to receive and send an email, browse a web page, access streaming media and the like on the terminal 600, and provides the user with wireless broadband Internet access.
  • FIG. 8 shows the WiFi module 170, it may be understood that the WiFi module 170 is not a mandatory component of the terminal 600, and may be removed as required without departing from the essence of the present application.
  • the processor 180 is a control center of the terminal 600, which connects parts of a whole mobile phone by using various interfaces and circuits, and executes various functions of the terminal 600 and processes data by operating or executing software programs and/or modules stored in the memory 120 and invoking data stored in the memory 120, so as to perform overall monitoring on the mobile phone.
  • the processor 180 may include one or more processing cores.
  • the processor 180 may be integrated with an application processor and a modem processor.
  • the application processor mainly processes an operating system, a user interface, an application, and the like.
  • the modem processor mainly processes wireless communication. It may be understood that the modem processor may also not be integrated into processor 180.
  • the terminal 600 further includes the power source 190 (for example, a battery) that provides power to the components.
  • the power source may be logically connected to the processor 180 by using a power source management system, to implement functions, such as charging, discharging, and power consumption management by using the power source management system.
  • the power source 190 may further include one or more of components, such as a direct current or an alternating current power source, a recharging system, a power source fault detecting circuit, a power source converter or inverter, and a power source state indicator.
  • the display unit of the terminal is a touch screen display.
  • the terminal further includes the memory and one or more programs, where the one or more programs are stored in the memory, and are executed by one or more processors after configuration.
  • the one or more programs include instructions for performing the following operations:
  • successively capturing video data for multiple time periods and determining, for each of the captured video data, whether a total length of time of the successively captured video data reaches a preset time length;
  • the memory of the terminal further includes instructions for performing the following operations:
  • the step of determining mentioned above includes determining, for the video data captured for each time period, whether the value of the timer is cleared; and when the value of the timer is cleared, determining that the total length of time that the video data was captured over the multiple time periods reaches the preset length of time.
  • the memory of the terminal further includes instructions for performing the following operation:
  • the memory of the terminal further includes instructions for performing the following operations:
  • the memory of the terminal further includes instructions for performing the following operations:
  • the terminal provided in the present application successively captures video data for multiple time periods, and combines the captured video data into one piece of video data.
  • one piece of video data is obtained by selectively capturing video data for multiple time periods, and content of captured video data can be selected at will. Therefore, the manner of obtaining video data is extended, and redundancy of obtained video data is decreased, thereby improving user experience.
  • the embodiment of the present application further provides a computer readable storage medium.
  • the computer readable storage medium may be the computer readable storage medium included in the memory in the foregoing embodiment, and may also be a computer readable storage medium that exists independently and is not assembled into terminal.
  • the computer readable storage medium stores one or more programs, where the one or more programs are used by one or more processors to execute the method for obtaining video data.
  • the method includes:
  • the method before the step of successively capturing video data for multiple times, the method further includes:
  • the above step of determining includes determining, each time video data is captured, whether the value of the timer is cleared; and when the value of the timer is cleared, determining that the total length of time video data is captured over the successive time periods reaches the preset time length.
  • the step of combining the video data into one piece of video data includes combining the video data captured over the successive time periods into one piece of video data according to a time sequence that the video data was captured.
  • the step of combining the video data into one piece of video data includes displaying identifiers corresponding to the video data captured during the multiple time periods, and combining the video data into one piece of video data according to how the user sorts the identifiers corresponding to the video data.
  • the method further includes:
  • the computer readable storage medium provided in the embodiment of the present application captures video data for multiple time periods, and combines the captured video data into one piece of video data.
  • one piece of video data may be obtained by selectively capturing video data for multiple time periods, and content of captured video data can be selected at will. Therefore, the manner of obtaining video data is extended, and redundancy of obtained video data is decreased, thereby improving user experience.
  • a further embodiment of the present application provides a graphic user interface, which is used on a terminal.
  • the terminal includes a touch screen display, a memory, and one or more processors for performing one or more programs.
  • the graphic user interface performs the following operations:
  • the graphic user interface provided in the embodiment of the present application successively captures video data for multiple time periods, and combines the video data obtained over the multiple time periods into one piece of video data.
  • one piece of video data is obtained by selectively capturing video data for multiple time periods, and content of captured video data can be selected at will. Therefore, the manner of obtaining video data is extended, and redundancy of obtained video data is decreased, thereby improving user experience.
  • division of the foregoing function modules is merely exemplary for describing how the apparatus for obtaining video data provided in the foregoing embodiments obtains video data.
  • functions may be allocated as required to different function modules for implementation. That is, the internal structure of the apparatus may be divided into different function modules, so as to perform all or a part of the functions described in the foregoing.
  • the embodiments of the apparatus for obtaining video data and the embodiments of the method for obtaining video data fall within a same conception. Reference may be made to the method embodiments for a specific implementation process of the apparatus for obtaining video data, which is not described in detail herein.
  • the term “if” may be construed to mean “when” or “upon” or “in response to determining” or “in accordance with a determination” or “in response to detecting, " that a stated condition precedent is true, depending on the context.
  • the phrase “if it is determined [that a stated condition precedent is true] “ or “if [a stated condition precedent is true] “ or “when [a stated condition precedent is true]” may be construed to mean “upon determining” or “in response to determining” or “in accordance with a determination” or “upon detecting” or “in response to detecting” that the stated condition precedent is true, depending on the context.
  • stages that are not order dependent may be reordered and other stages may be combined or broken out. While some reordering or other groupings are specifically mentioned, others will be obvious to those of ordinary skill in the art and so do not present an exhaustive list of alternatives. Moreover, it should be recognized that the stages could be implemented in hardware, firmware, software or any combination thereof.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Studio Devices (AREA)
  • Information Transfer Between Computers (AREA)

Abstract

A method of obtaining and sharing video data using a social networking application includes at a computer having one or more processors and memory storing programs executed by the one or more processors, recording a plurality of video segments successively; marking each video segment with a respective identifier; caching the plurality of video segments in the memory for processing; determining whether a total time length of the plurality of video segments exceeds a pre-set threshold; stopping video recording if the total time length of the plurality of video segments exceeds the pre-set threshold; and integrating the plurality of video segments into one video according to a predefined order.

Description

METHOD, APPARATUS, AND TERMINAL FOR OBTAINING VIDEO DATA
RELATED APPLICATION
This application claims priority to Chinese Patent Application No. 201310452594.5, “METHOD, APPARATUS, AND TERMINAL FOR OBTAINING VIDEO DATA, ” filed on September 27, 2013, which is hereby incorporated by reference in its entirety.
TECHNICAL FIELD
The present application relates to the field of multimedia technologies, and in particular, to a method, an apparatus, and a terminal for obtaining video data.
BACKGROUND
With the development of the modern society, as basic tools for people, mobile terminals become increasingly popular. For these terminals, many corresponding applications are available for installation and use. Among the applications, a large number of applications have a video data obtaining function. In reality, users of the video data obtaining function expect that the content of obtained video data is what they wish to display, and further, the length of the video data can be as short as possible. Therefore, selecting a reasonable method for obtaining video data enables a user that uses a related function to have better user experience.
In the prior art, a method for obtaining video data includes triggering a camera apparatus on a terminal to successively capture one piece of video data by clicking a related interface interaction element on an application interface, for example, a button or a touch control option for capturing a video; and using the captured consecutive video data as the obtained video data.
In a process of implementing the present application, the inventor discovers that the prior art has at least some problems as follows. 
Because the capturing process is consecutive, the obtained video data is captured by the camera apparatus at one time, and thus, the video capturing manner is monotonous. In addition, while capturing video in a moving status, data that is not required by a user is often captured. As a result, the obtained video data includes redundant video data, and the redundancy of the obtained video data is increased.
SUMMARY
The above deficiencies and other problems associated with the conventional approach of obtaining and sharing video data are reduced or eliminated by the present application disclosed below. In some embodiments, the present application is implemented in a system of obtaining and sharing video data using a social networking application that has one or more processors, memory and one or more modules, programs or sets of instructions stored in the memory for performing multiple functions. Instructions for performing these functions may be included in a computer program product configured for execution by one or more processors.
One aspect of the present application involves a computer-implemented method of obtaining and sharing video data using a social networking application performed by a computer having one or more processors and memory. The computer-implemented method includes: recording a plurality of video segments successively; marking each video segment with a respective identifier; caching the plurality of video segments in the memory for processing; determining whether a total time length of the plurality of video segments exceeds a pre-set threshold; stopping video recording if the total time length of the plurality of video segments exceeds the pre-set threshold; and integrating the plurality of video segments into one video according to a predefined order.
In some embodiments, determining whether the total time length of the plurality of video segments exceeds the pre-set threshold further comprises configuring a timer with a value equal to the pre-set threshold; subtracting time length of the video segment from the value of the timer once the video segment is recorded; determining whether the value of the timer equals to or less than zero; and when the value of the timer equals to or less than zero, instructing that the total time length of the plurality of video segments exceeds the pre-set threshold.
In some embodiments, integrating the plurality of video segments into the one video according to a predefined order further comprises reading the plurality of identifiers that correspond to the plurality of video segments, respectively; re-organizing the plurality of video segments into a sequence based on sorting of the plurality of identifiers; and aggregating the re-organized plurality of video segments into the one video.
In some embodiments, integrating the plurality of video segments into the one video according to a predefined order further comprises reading a plurality of recording times that correspond to the plurality of video segments, respectively; re-organizing the plurality of video segments into a sequence based on sorting of the plurality of recording times; and aggregating the re-organized plurality of video segments into the one video.
In some embodiments, the method of obtaining and sharing video data using a social networking application further comprises prompting a message to a user of the social networking application whether to share the one video in a social network; and if receiving an instruction from the user to share the one video in the social network, choosing a participant of the social network that the user wants to share with, and sending the one video to the participant via the social network application.
Another aspect of the present application involves a system of obtaining and sharing video data using a social networking application. The system includes memory, one or more processors, and one or more programs stored in the memory and configured for execution by the one or more processors. The one or more programs include: recording a plurality of video segments successively; marking each video segment with a respective identifier; caching the plurality of video segments in the memory for processing; determining whether a total time length of the plurality of video segments exceeds a pre-set threshold; stopping video recording if the total time length of the plurality of video segments exceeds the pre-set threshold; and integrating the plurality of video segments into one video according to a predefined order.
Another aspect of the present application involves a non-transitory computer readable storage medium having stored therein one or more programs for execution by one or more processors of a system of obtaining and sharing video data using a social networking application, the one or more programs including instructions for: recording a plurality of  video segments successively; marking each video segment with a respective identifier; caching the plurality of video segments in the memory for processing; determining whether a total time length of the plurality of video segments exceeds a pre-set threshold; stopping video recording if the total time length of the plurality of video segments exceeds the pre-set threshold; and integrating the plurality of video segments into one video according to a predefined order.
BRIEF DESCRIPTION OF THE DRAWINGS
The aforementioned features and advantages of the present application as well as additional features and advantages thereof will be more clearly understood hereinafter as a result of a detailed description of preferred embodiments when taken in conjunction with the drawings.
To describe the technical solutions in the embodiments of the present application more clearly, the following briefly introduces the accompanying drawings required for describing the embodiments. Apparently, the accompanying drawings in the following description show merely some embodiments of the present application, and a person of ordinary skill in the art may still derive other drawings from these accompanying drawings without creative efforts.
FIG. 1 is a flowchart of a method for obtaining video data according to Embodiment 1 of the present application;
FIG. 2 is a flowchart of a method for obtaining video data according to Embodiment 2 of the present application;
FIG. 3 is a first sequence diagram of identifiers corresponding to video data according to Embodiment 2 of the present application;
FIG. 4 is a second sequence diagram of identifiers corresponding to video data according to Embodiment 2 of the present application;
FIG. 5 is a first schematic structural diagram of an apparatus for obtaining video data according to Embodiment 3 of the present application; 
FIG. 6 is a second schematic structural diagram of the apparatus for obtaining video data according to Embodiment 3 of the present application;
FIG. 7 is a third schematic structural diagram of the apparatus for obtaining video data according to Embodiment 3 of the present application; and
FIG. 8 is a schematic structural diagram of a terminal according to Embodiment 4 of the present application.
Like reference numerals refer to corresponding parts throughout the several views of the drawings.
DESCRIPTION OF EMBODIMENTS
Reference will now be made in detail to embodiments, examples of which are illustrated in the accompanying drawings. In the following detailed description, numerous specific details are set forth in order to provide a thorough understanding of the subject matter presented herein. But it will be apparent to one skilled in the art that the subject matter may be practiced without these specific details. In other instances, well-known methods, procedures, components, and circuits have not been described in detail so as not to unnecessarily obscure aspects of the embodiments.
To make the objectives, technical solutions, and advantages of the present application more comprehensible, the following further describes the embodiments of the present application in detail with reference to the accompanying drawings.
Embodiment 1
The embodiment of the present application provides a method for obtaining video data. The method is applicable to a social network application and installed on a terminal with a capturing function. The terminal includes, but is not limited to, a mobile phone, a computer, a tablet, and the like. A specific form of the terminal is not limited in this embodiment. Referring to FIG. 1, the method provided in this embodiment includes: 
t step 101, a plurality of video segments is recorded successively, each video segment is marked with a respective identifier, and then it is determined whether a total time length of the plurality of video segments exceeds a pre-set threshold.
In some embodiments, before the step of successively recording a plurality of video segments, the following steps may be performed:
configuring a timer with a value equal to the pre-set threshold;
subtracting time length of the video segment from the value of the timer once the video segment is recorded;
determining whether the value of the timer equals to or less than zero; and
when the value of the timer equals to or less than zero, instructing that the total time length of the plurality of video segments exceeds the pre-set threshold.
At step 102, the video recording is stopped if the total time length of the plurality of video segments exceeds the pre-set threshold. In addition, the plurality of video segments are integrated into one video according to a predefined order.
In some embodiments, the step of integrating the plurality of video segments into one video according to a predefined order includes:
reading the plurality of identifiers that correspond to the plurality of video segments, respectively;
re-organizing the plurality of video segments into a sequence based on sorting of the plurality of identifiers; and
aggregating the re-organized plurality of video segments into the one video.
Yet in some embodiments, the step of integrating the plurality of video segments into one video according to a predefined order includes:
reading a plurality of recording times that correspond to the plurality of video segments, respectively;
re-organizing the plurality of video segments into a sequence based on sorting of the plurality of recording times; and
aggregating the re-organized plurality of video segments into the one video.
In some embodiments, after the step of integrating the plurality of video segments into one video according to a predefined order, the following steps may be performed:
prompting whether to share the video data obtained after combination; and
if receiving an indication about sharing the video data obtained after combination, transmitting the video data obtained after combination to another terminal.
In the method provided in this embodiment, video data is successively captured for multiple time periods, and the video data obtained by successively capturing for the multiple time periods is combined into one piece of video data. In this way, one piece of video data is obtained by selectively capturing for multiple time periods, and content of captured video data can be selected at will. Therefore, the manner of obtaining video data is extended, and redundancy of obtained video data is decreased, thereby improving user experience.
Embodiment 2
With reference to the content in Embodiment 2, the embodiment of the present application provides a method for obtaining video data. The method is applicable to a social network application and installed in a terminal with a capturing function. The terminal includes, but is not limited to, a mobile phone, a computer, a tablet, and the like. A specific form of the terminal is not limited in this embodiment. Referring to FIG. 2, a procedure of the method provided in this embodiment includes:
At step 201, successively capturing video data for multiple time periods, and determining, each time period for which video data is captured, whether a total time length of the video data obtained by successively capturing for the multiple times reaches a preset time length.
Specifically, before video data is captured, the preset time length of capturing video data may be set in advance, for example, by a provider that provides a video data obtaining service. Alternatively, if video data is obtained in a social network application, a corresponding operation interface may be provided in the social network application, so that a user of the social network application can set the preset time length by using the operation  interface. The time length set by the user is used as a determined preset time length of capturing video data. In a practical operation, the preset time length of capturing video data may also be set by selecting other methods, which is not specifically limited in this embodiment. Besides, the determined preset time length is also not limited in this embodiment, and includes, but is not limited to, 6 seconds, 8 seconds, or the like.
No matter which manner is used and how long the preset time length is, after the preset time length is determined, a timer may be set according to the determined preset time length. An initial value of the timer may be the determined preset time length. When video data is successively captured for multiple times in the application corresponding to the social network, the value of the timer is subtracted by a time length of video data obtained each time video data is captured, until the value of the timer is subtracted to zero. Then, it may be determined whether the total time length of the video data obtained by successively capturing for the multiple times reaches the preset time length.
Therefore, before the step of successively capturing video data for multiple times, the following steps may be performed:
setting the preset time length on the timer, where the value of the timer decreases progressively according to a time length of capturing video data, and is cleared when the total time length of capturing video data reaches the preset time length; and
the step of determining, for each time capturing video data, whether a total time length of the video data obtained by successively capturing for the multiple times reaches a preset time length includes:
determining, for each time capturing video data, whether the value of the timer is cleared; and when the value of the timer is cleared, determining that the total time length of the video data obtained by successively capturing for the multiple times reaches the preset time length.
A case in which the determined preset time length is 6 seconds (s) is used as an example. In this case, the initial value of the timer may be set to 6 sec. Next, capturing video data is successively performed multiple times. If a time length of capturing video data for the first time is 2.5 sec, the value of the timer is subtracted by 2.5 sec. At this point, the value of the timer is 3.5. If a time length of capturing video data for the second time is 2 sec,  the value of the timer is subtracted by 2 sec. At this time, the value of the timer is 1.5 sec, which means that a remained capturing time length is 1.5 sec. When video data is captured again and a time length reaches 1.5 sec, the value of the timer is subtracted by 1.5 sec. At this time, the value of the timer is 0, i. e. , the timer is cleared. In this case, it may be determined that a total time length of video data obtained by successively capturing for three times reaches the preset time length. Therefore, capturing of video data is stopped.
If a time length of capturing video data is not limited, the preset time length may be set to a larger value, so that capturing can be successively performed for multiple times to obtain complete video data expected by a user, and capturing is not stopped in the capturing process when the preset time length is reached. In addition, multiple manners are available for triggering capturing of video data each time, and may include, but is not limited to, detecting an operation of pressing a capturing button. If it is detected that the capturing button is pressed, a signal of capturing video data is obtained, and capturing of video data is performed accordingly. Also, multiple manners are available for stopping capturing of video data each time, and may include, but is not limited to, detecting an operation of pressing a capturing stopping button. If it is detected that the capturing stopping button is pressed, a signal of stopping current capturing of video data is obtained, and capturing of video data is stopped. The capturing button and the capturing stopping button may be two buttons set on the terminal. In addition, there may be other manners of triggering capturing of video data and of triggering stopping capturing video data, which are not specifically limited in this embodiment.
At step 202, if the total time length of the video data obtained by successively capturing for the multiple times reaches the preset time length, capturing of video data is stopped.
Specifically, when it is determined that the total time length of the video data obtained by successively capturing for the multiple times reaches the preset time length, the application corresponding to the social network may stop capturing of video data. When stopping capturing of video data, the application corresponding to the social network may display related prompt information to notify the user that capturing of video data ends. For example, a dialog box may be displayed on the interface of the social network application,  and information displayed in the dialog box is used to notify the user that capturing of video data ends. Other manners may also be selected to prompt that capturing of video data ends. The specific prompt manner is not limited in this embodiment.
Further, when the total time length of the video data is close to the preset time length, in the method provided in this embodiment, corresponding prompt information may further be displayed to prompt the user that the capturing process will end soon. A manner of displaying the prompt information and content of the displayed prompt information are also not specifically limited in this embodiment.
It should be noted that in some embodiments, when the total time length of the video data obtained by successively capturing for the multiple times reaches the preset time length, capturing of video data is stopped, so as to obtain one piece of video data in a subsequent step. Yet in some other embodiments, when it is applied to the social network application, the method may further provide a related operation interface through which the user can trigger an operation to stop capturing of video data. Specifically, a manner of providing the related operation interface by the social network application is not specifically limited in this embodiment, and includes, but is not limited to, displaying a corresponding operation option on an interaction interface provided by the application corresponding to the social network, and determining, by detecting a selection operation of the user on the operation option, whether to stop capturing of video data.
The interaction interface is a channel for information exchange between the user and the terminal (a computer, a notebook, a mobile terminal, a mobile phone, or the like) . The user uses the interaction interface to input information on the terminal and perform an operation. The terminal uses the interaction interface to provide information to the user for reading, analyzing, and determining by the user. The interaction interface in this embodiment includes a home page displayed when software or an application is opened, and interaction interfaces entered after various function buttons and entries are clicked. Interaction interfaces that appear at different levels or due to different triggering are different. Interaction interfaces entered due to a same triggering or from a same entry are the same regardless of whether reading content on the interfaces changes. Different interfaces generally mean that  interaction interfaces at a certain level and displayed to a user are different in the interaction interfaces themselves, and do not mean that the interaction interfaces are different in content.
For example, a "combine" option is displayed on an interaction interface provided by the application corresponding to the social network. After obtaining two pieces of video data by successively capturing video data for two times, where a total time length of the two pieces of video data has not reached a preset time length yet, if it is detected that the user selects the "combine" button by a clicking operation, an indication of stopping capturing video data is obtained, and the two pieces of video data obtained by capturing before is used as video data obtained by successively capturing for multiple times, and a combining operation is performed in a subsequent step.
In addition to the manner of displaying a corresponding operation option on the interaction interface provided by the social network application, other manners may also be used, which are not specifically limited in this embodiment.
At step 203, the video data obtained by successively capturing for the multiple times is combined into one piece of video data, and use the video data obtained after combination as obtained video data.
In this step, the step of combining the video data obtained by successively capturing for the multiple times into one piece of video data includes, but is not limited to, the following two manners:
A first combining technique may include combining, according to a time sequence of capturing video data, the video data obtained by successively capturing for the multiple times into one piece of video data.
In the first combining technique, multiple different pieces of video data are obtained after successively capturing for multiple times. Because different pieces of video data are specifically captured at different times in a capturing process, the multiple pieces of video data obtained after successively capturing for the multiple times are sorted according to the specific capturing times. Then, the sorted multiple pieces of video data are combined into one piece of video data, and the video data obtained after combination is used as obtained video data. The manner of sorting, according to the specific capturing times, the multiple  pieces of video data obtained after successively capturing for the multiple times is not specifically limited in this embodiment. For example, the multiple pieces of video data may be sorted according to the capturing times in an early-to-late sequence, or in a late-to-early sequence.
Specifically, when the multiple pieces of video data are combined into one piece of video data, the multiple pieces of video data may be combined according to specific file formats of the video data. The method of combining the video data obtained by successively capturing for the multiple times into one piece of video data is not specifically limited in this embodiment.
For example, three pieces of video data are obtained after successively capturing for three times. A first piece is video data V_1 of 2.5 sec, a second piece is video data V_2 of 2 sec, and a third piece is video data V_3 of 1.5 sec. Specific capturing times of the three pieces of video data are respectively 10: 31 am, 10: 27 am, and 10: 40 am in a same day. Accordingly, the three pieces of video data are sorted according to the specific capturing times in an early-to-late sequence, and a final sorted sequence is V_2, V_1, and V_3. Next, the three pieces of video data may be combined in the foregoing sequence, and video data obtained after combination is used as obtained video data.
A second combining technique may include displaying identifiers corresponding to the video data obtained by successively capturing for the multiple times, and combining the video data obtained by successively capturing for the multiple times into one piece of video data according to how the user sorts the identifiers corresponding to the video data obtained by successively capturing for the multiple times.
Specifically, for each time a piece of video data is captured, an identifier may be set for the piece of video data. Different pieces of video data correspond to different identifiers. After video data is successively captured for multiple times, identifiers corresponding to video data obtained by successively capturing for the multiple times may be displayed on an interface. A social network application may provide a related operation interface. A user may sort, by using the related operation interface, the identifiers corresponding to the video data obtained by successively capturing for the multiple times. Finally, the social network application combines, according to how the user sorts the  identifiers, the video data obtained by successively capturing for the multiple times into one piece of video data, and uses video data obtained after combination as obtained video data.
For example, a user obtains three pieces of video data after successively capturing video three times with a touch screen mobile phone. A first piece is video data V_1 of 2.5 sec, a second piece is video data V_2 of 2 sec, and a third piece is video data V_3 of 1.5 sec. Identifiers of the three pieces of video data are respectively V_1, V_2, and V_3. The identifiers on an interface and corresponding to the video data obtained by successively capturing video three times are shown in FIG. 3. The user may sort icons corresponding to the three identifiers displayed on the interface shown in FIG. 3, so that the identifiers corresponding to the video data obtained by successively capturing for three times are sorted. A final sorting result is shown in FIG. 4. It can be known from FIG. 4 that the sorting result is V_3, V_1, and V_2. Then, a first application may combine, according to the foregoing sequence, the video data, V_3, V_2, and V_1, obtained by successively capturing for three times into one piece of video data.
Further, after the step of combining the video data obtained by successively capturing video multiple times into one piece of video data, the following steps may be performed:
prompting whether to share the video data obtained after the combination; and
if an indication to share the combined video data is obtained, transmitting, by using an application corresponding to a social network, the combined video data to another terminal.
Specifically, after obtaining the combined video data, the social network application may issue prompt information to prompt a user whether to share the combined video data, for example, a dialog box or a pop-up window may be used, which is not specifically limited in this embodiment. After receiving the prompt information, the user may select according to the prompt information whether to share the combined video data. The indication about sharing the combined video data is obtained by detecting a selection result of the user, and a sharing operation is performed according to the indication. A case in which the method provided in this embodiment is applied to a social network application is still used as an example. After obtaining the indication about sharing the combined video  data, the social network application may transmit the combined video data to a corresponding terminal according to sharing destination information included in the indication. Specific content of the sharing destination information included in the indication is not limited in this embodiment, and includes, but is not limited to, a user nickname, an account number, and the like registered by the user in the application corresponding to the social network. Objects that shares others' video data and whose video data is shared by others may be users registered in applications corresponding to a same social network, and may also be users registered in applications corresponding to different social networks. That is, the combined video data obtained may be shared on an application platform corresponding to a same social network, or on application platforms corresponding to different social networks. A specific sharing manner and scenario are not limited in this embodiment.
For example, a terminal A is associated with an account number A1 registered by a user A in an application corresponding to a first social network, a terminal B is associated with an account number B registered by a user B in the application corresponding to the first social network, the terminal A is further associated with an account number A2 registered by the user A in an application corresponding to a second social network, and a terminal C is associated with an account number C registered by a user C in the application corresponding to the second social network; and an identifier of user A in the application corresponding to the first social network is A1, an identifier of user B in the application corresponding to the first social network is B, an identifier of user A in the application corresponding to the second social network is A2, and an identifier of user C in the application corresponding to the second social network is C.
After obtaining the combined video data in the application corresponding to the first social network; terminal A detects that user A selects the identifier B corresponding to user B on an interaction interface provided by the application corresponding to the first social network; and detects that user A selects a sharing option by using an interaction interface provided by the application corresponding to the social network, terminal A then determines that the combined video data needs to be shared with a terminal corresponding to identifier B. Therefore, terminal A sends the combined video data and identifier B to a server of the application corresponding to the first social network. After finding that the terminal corresponding to identifier B is terminal B, the server forwards, to terminal B, the combined  video data sent by terminal A. In this way, the combined video data obtained by terminal A is shared with user B corresponding to terminal B.
In another embodiment, after terminal A obtains the combined video data in the application corresponding to the first social network, and detects that user A selects the sharing option on the interaction interface provided by the application corresponding to the first social network. Terminal A then displays, on the interaction interface provided by the application corresponding to the social network, applications with which the video data can be shared. The applications with which the video data can be shared include the application corresponding to the second social network. After detecting that user A selects the application corresponding to the second social network, terminal A displays a friend list of user A in the application corresponding to the second social network. User C is displayed in the friend list. After detecting that user A selects identifier C corresponding to user C, terminal A sends the video data obtained after combination and identifier C to a server of the application corresponding to the second social network. After finding that a terminal corresponding to identifier C is terminal C, the server forwards, to terminal C, the combined video data sent by terminal A. In this way, the video data obtained by terminal A is shared with user C corresponding to terminal C.
It should be noted that the two examples used in the foregoing are merely exemplary for describing the sharing process. Besides the foregoing two sharing manners, other sharing manners may also be used, which are not specifically limited in this embodiment.
In the method provided in this embodiment, video data is successively captured multiple times, and the captured video data is combined into one piece of video data. In this way, one piece of video data is obtained by selectively capturing video multiple times, and the content of captured video data can be selected at will. Therefore, the manner of obtaining video data is extended, and redundancy of obtained video data is decreased, thereby improving user experience.
Embodiment 3
Referring to FIG. 5, an embodiment of the present application provides an apparatus for obtaining video data. The apparatus includes:
capturing module 501, configured to successively capture video data multiple times;
a determining module 502, configured to determine, each time video data is captured, whether a total time length of the video data obtained by successively capturing for the multiple times reaches a preset time length;
control module 503, configured to control the capturing module to stop capturing of video data if the total time length of the video data obtained by successively capturing for the multiple times reaches the preset time length; and
a combining module 504, configured to combine the video data obtained by successively capturing for the multiple times into one piece of video data, and use video data obtained after combination as obtained video data.
In some embodiments, referring to FIG. 6, the apparatus for obtaining video data further includes:
setting module 505, configured to set the preset time length on a timer, where a value of the timer decreases progressively according to a time length of capturing video data, and is cleared when the total time length of capturing video data reaches the preset time length; and
the determining module 502 is configured to determine, each time video data is captured, whether the value of the timer is cleared; and when the value of the timer is cleared, determine that the total time length of video data obtained by successively capturing for the multiple times reaches the preset time length.
In some embodiments, the combining module 504 is configured to display identifiers corresponding to the video data obtained by successively capturing for the multiple times, and combine the video data into one piece of video data according to how a user sorts the identifiers corresponding to the video data obtained by successively capturing for the multiple times.
In some embodiments, referring to FIG. 7, the apparatus for obtaining video data further includes:
prompt module 506, configured to prompt whether to share the video data obtained after combination; and
transmitting module 507, configured to transmit the video data obtained after combination to another terminal after an indication about sharing the combined video data is received.
The apparatus provided in this embodiment successively captures video data for multiple times, and combines the video data obtained by successively capturing for the multiple times into one piece of video data. In this way, one piece of video data is obtained by selectively capturing for multiple times, and content of captured video data can be selected at will. Therefore, the manner of obtaining video data is extended, and redundancy of obtained video data is decreased, thereby improving user experience.
Embodiment 4
This embodiment provides a terminal. The terminal may be configured to execute the method for obtaining video data provided in the foregoing embodiments.
Referring to FIG. 8, the terminal 600 may include components such as a radio frequency (RF) circuit 110, a memory 120 including one or more computer readable storage media, an input unit 130, a display unit 140, a sensor 150, an audio circuit 160, a wireless fidelity (WiFi) module 170, a processor 180 including one or more processing cores, and a power source 190. A person skilled in the art can understand that the terminal structure shown in FIG. 8 does not constitute a limitation to the terminal. The terminal may include different components, additional components, or fewer components, or any combination of the foregoing than shown in FIG. 8.
The RF circuit 110 may be configured to receive and send a signal during communication, or information receiving and sending. In particular, after receiving downlink information of a base station, the RF circuit 110 sends the downlink information to one or more processors 180 for processing. In addition, the RF circuit 110 sends uplink data to the base station. Generally, the RF circuit 110 includes, but is not limited to, an antenna, at least one amplifier, a tuner, one or more oscillator, a subscriber identity module (SIM) card, a transceiver, a coupler, a low noise amplifier (LNA) , a duplexer, and the like. The RF circuit  110 may further communicate with other devices through a network by means of wireless communication. Any communication standard or protocol may be used for the wireless communication, including but not limited to Global System for Mobile Communications (GSM) , General Packet Radio Service (GPRS) , Code Division Multiple Access (CDMA) , Wideband Code Division Multiple Access (WCDMA) , Long Term Evolution (LTE) , email, Short Messaging Service (SMS) , and the like.
The memory 120 may be configured to store a software program and module. The processor 180 executes function application and data processing by operating the software program and module stored in the memory 120. The memory 120 may mainly include a program storing area and a data storing area. The program storing area may store an operating system, at least one application required by functions (for example, an audio playing function and an image playing function) , and the like. The data storing area may store data (for example, audio data and a phonebook) that is created according to the use of the terminal 600 and the like. The memory 120 may also include a high-speed random access memory, and may further include a non-volatile memory, for example, include at least one disk storage device, a flash memory, or other volatile solid-state storage devices. Correspondingly, the memory 120 may further include a memory controller to control access of the processor 180 and the input unit 130 to the memory 120.
The input unit 130 may be configured to receive inputted numerical or character information and generate keyboard, mouse, joystick, optical, and/or trackball signal input related to user setting and function control. Specifically, the input unit 130 may include a touch-sensitive surface 131 and a secondary input device 132. The touch-sensitive surface 131 is also referred to as a touch display screen or a touch pad. The touch-sensitive surface 131 may collect a touch operation performed by a user on or near the touch-sensitive surface 131 (for example, an operation performed by a user with any proper object or attachment, such as a finger or a touch pen, on or near the touch-sensitive surface 131) , and drive a corresponding connecting apparatus according to a preset program. Optionally, the touch-sensitive surface 131 may include two parts, a touch detecting apparatus and a touch controller. The touch detecting apparatus detects a touch position of a user and a signal brought by a touch operation, and transmits the signal to the touch controller. The touch controller receives touch information from the touch detecting apparatus, converts the touch  information to a touch point coordinate, and sends the touch point coordinate to the processor 180. The touch controller can receive and execute a command sent by the processor 180. Moreover, the touch-sensitive surface 131 may be implemented in a resistive, capacitive, infrared, or a surface acoustic form, for example. Besides the touch-sensitive surface 131, the input unit 130 may further include a secondary input device 132. Specifically, the secondary input device 132 may include, but is not limited to, one or more of a physical keyboard, a functional key (for example, a volume control key or a switch key) , a trackball, a mouse, a joystick and the like.
The display unit 140 may be configured to display information input by a user, information provided to the user, or various graphic user interfaces of the terminal 600. The graphic user interfaces may include a graph, a text, an icon, a video, and any combination thereof. The display unit 140 may include a display panel 141. Optionally, the display panel 141 may be configured in a form of a liquid crystal display (LCD) , an organic light-emitting diode (OLED) , or the like. Further, the touch-sensitive surface 131 may cover the display panel 141. After detecting a touch operation performed on or near the touch-sensitive surface 131, the touch-sensitive surface 131 transmits the touch operation to the processor 180 to determine a type of a touch event. Then, the processor 180 provides corresponding visual output on the display panel 141 according to the type of the touch event. Though in FIG. 8, the touch-sensitive surface 131 and the display panel 141 implement input and output functions as two independent components, in some embodiments, the touch-sensitive surface 131 and the display panel 141 may be integrated to implement input and output functions.
The terminal 600 may further include at least one sensor 150, for example, a light sensor, a motion sensor, and other sensors. Specifically, the light sensor may include an ambient light sensor and a proximity sensor. The ambient light sensor may adjust brightness of the display panel 141 according to brightness of ambient light. The proximity sensor may turn off the display panel 141 and/or backlight when the terminal 600 moves near an ear of a user. As a motion sensor, a gravity acceleration sensor may detect magnitudes of accelerations of all directions (generally, three axes) , and can detect a magnitude and direction of gravity in a stationary state. The gravity acceleration sensor may be configured for applications of recognizing a mobile phone posture (for example, switch between portrait and landscape screens, related games, magnetometer calibration) , related functions of  vibration recognition (for example, a pedometer and knocking) , and the like. The terminal 600 may further be configured with other sensors, such as a gyroscope, a barometer, a hygrometer, a thermometer, and an infrared sensor, which are not described in detail herein.
The audio circuit 160, a loudspeaker 161, a microphone 162 may provide an audio interface between a user and the terminal 600. The audio circuit 160 may transmit an electrical signal obtained by converting received audio data to the loudspeaker 161, so that the loudspeaker 161 converts the electrical signal to a sound signal for output. The microphone 162 converts a collected sound signal to an electrical signal. The audio circuit 160 receives the electrical signal, converts the electrical signal to audio data, and outputs the audio data to the processor 180 for processing. After processing, the audio data is sent through the RF circuit 110 to, for example, another terminal, or is output to the memory 120 for further processing. The audio circuit 160 may further include an earphone jack to provide communication between an external earphone and the terminal 600.
WiFi is a short-distance wireless transmission technology. The WiFi module 170 enables a user to receive and send an email, browse a web page, access streaming media and the like on the terminal 600, and provides the user with wireless broadband Internet access. Though FIG. 8 shows the WiFi module 170, it may be understood that the WiFi module 170 is not a mandatory component of the terminal 600, and may be removed as required without departing from the essence of the present application.
The processor 180 is a control center of the terminal 600, which connects parts of a whole mobile phone by using various interfaces and circuits, and executes various functions of the terminal 600 and processes data by operating or executing software programs and/or modules stored in the memory 120 and invoking data stored in the memory 120, so as to perform overall monitoring on the mobile phone. Optionally, the processor 180 may include one or more processing cores. Preferably, the processor 180 may be integrated with an application processor and a modem processor. The application processor mainly processes an operating system, a user interface, an application, and the like. The modem processor mainly processes wireless communication. It may be understood that the modem processor may also not be integrated into processor 180.
The terminal 600 further includes the power source 190 (for example, a battery) that provides power to the components. Preferably, the power source may be logically connected to the processor 180 by using a power source management system, to implement functions, such as charging, discharging, and power consumption management by using the power source management system. The power source 190 may further include one or more of components, such as a direct current or an alternating current power source, a recharging system, a power source fault detecting circuit, a power source converter or inverter, and a power source state indicator.
Though not shown in the diagram, a camera, a blue tooth module, and the like may further be included in the terminal 600, which are not described in detail herein. Specifically, in this embodiment, the display unit of the terminal is a touch screen display. The terminal further includes the memory and one or more programs, where the one or more programs are stored in the memory, and are executed by one or more processors after configuration. The one or more programs include instructions for performing the following operations:
successively capturing video data for multiple time periods, and determining, for each of the captured video data, whether a total length of time of the successively captured video data reaches a preset time length;
determining if the total length of time of the successively captured video data reaches the preset time length, and, if so, stopping video data capture;
combining the captured video data for the multiple time periods into one piece of video data; and
using the combined video data.
Assuming that the foregoing is a first possible implementation manner, based on the first possible implementation manner, in a second possible implementation manner, the memory of the terminal further includes instructions for performing the following operations:
setting the preset time length on a timer, where a value of the timer decreases progressively according to a length of time that video data is captured, and is cleared when the total length of time that video data is captured reaches the preset time length; and
the step of determining mentioned above, includes determining, for the video data captured for each time period, whether the value of the timer is cleared; and when the  value of the timer is cleared, determining that the total length of time that the video data was captured over the multiple time periods reaches the preset length of time.
Based on the first possible implementation manner, in a third possible implementation manner, the memory of the terminal further includes instructions for performing the following operation:
combining the video data obtained over the multiple time periods according to the time sequence that the video data was captured so as to obtain one piece of video data.
Based on the first possible implementation manner, in a fourth possible implementation manner, the memory of the terminal further includes instructions for performing the following operations:
displaying identifiers corresponding to the video data captured for the multiple time periods, and combining the captured video data into one piece of video data according to how the user sorts the identifiers corresponding to the captured video data.
Based on any one of the first to the fourth possible implementation manners, in a fifth possible implementation manner, the memory of the terminal further includes instructions for performing the following operations:
prompting whether to share the combined video data; and
if obtaining an indication about sharing the combined video data, transmitting the combined video data to another terminal.
The terminal provided in the present application successively captures video data for multiple time periods, and combines the captured video data into one piece of video data. In this way, one piece of video data is obtained by selectively capturing video data for multiple time periods, and content of captured video data can be selected at will. Therefore, the manner of obtaining video data is extended, and redundancy of obtained video data is decreased, thereby improving user experience.
Embodiment 5
The embodiment of the present application further provides a computer readable storage medium. The computer readable storage medium may be the computer readable storage medium included in the memory in the foregoing embodiment, and may also  be a computer readable storage medium that exists independently and is not assembled into terminal. The computer readable storage medium stores one or more programs, where the one or more programs are used by one or more processors to execute the method for obtaining video data. The method includes:
successively capturing video data for multiple time periods, and determining, for each time period in which video data is captured, whether a total length of time that video data is captured reaches a preset time length;
determining if the total length of time for capturing the video data over the successive time periods reaches the preset time length, and, if so, stopping the capturing of video data;
combining the video data captured for the time periods into one piece of video data; and
using the combined video data obtained.
In a second possible implementation manner, before the step of successively capturing video data for multiple times, the method further includes:
setting the preset time length on a timer, where a value of the timer decreases progressively according to a length of time that video data is captured, and the timer is cleared when the total length of time that video data is captured reaches the preset time length; and
the above step of determining includes determining, each time video data is captured, whether the value of the timer is cleared; and when the value of the timer is cleared, determining that the total length of time video data is captured over the successive time periods reaches the preset time length.
Based on the first possible implementation manner, in a third possible implementation manner, the step of combining the video data into one piece of video data includes combining the video data captured over the successive time periods into one piece of video data according to a time sequence that the video data was captured.
Based on the first possible implementation manner, in a fourth possible implementation manner, the step of combining the video data into one piece of video data includes displaying identifiers corresponding to the video data captured during the multiple time periods, and combining the video data into one piece of video data according to how the user sorts the identifiers corresponding to the video data. 
Based on any one of the first to the fourth possible implementation manners, in a fifth possible implementation manner, after the step of combining the video data captured over the multiple time periods into one piece of video data, the method further includes:
prompting whether to share the combined video data; and
if receiving an indication about sharing the combined video data, transmitting the combined video data to another terminal.
The computer readable storage medium provided in the embodiment of the present application captures video data for multiple time periods, and combines the captured video data into one piece of video data. In this way, one piece of video data may be obtained by selectively capturing video data for multiple time periods, and content of captured video data can be selected at will. Therefore, the manner of obtaining video data is extended, and redundancy of obtained video data is decreased, thereby improving user experience.
Embodiment 6
A further embodiment of the present application provides a graphic user interface, which is used on a terminal. The terminal includes a touch screen display, a memory, and one or more processors for performing one or more programs. The graphic user interface performs the following operations:
successively capturing video data for multiple time periods, and determining, each time video data is captured, whether a total length of time of the captured video data for the multiple time periods reaches a preset time length; and
if the total length of time of the captured video data reaches the preset time length, stopping the capture of video data; combining the video data obtained for the multiple time periods into one piece of video data; and using the combined video data as obtained video data.
The graphic user interface provided in the embodiment of the present application successively captures video data for multiple time periods, and combines the video data obtained over the multiple time periods into one piece of video data. In this way, one piece of video data is obtained by selectively capturing video data for multiple time periods, and content of captured video data can be selected at will. Therefore, the manner of  obtaining video data is extended, and redundancy of obtained video data is decreased, thereby improving user experience.
It should be noted that division of the foregoing function modules is merely exemplary for describing how the apparatus for obtaining video data provided in the foregoing embodiments obtains video data. In a practical application, functions may be allocated as required to different function modules for implementation. That is, the internal structure of the apparatus may be divided into different function modules, so as to perform all or a part of the functions described in the foregoing. In addition, the embodiments of the apparatus for obtaining video data and the embodiments of the method for obtaining video data fall within a same conception. Reference may be made to the method embodiments for a specific implementation process of the apparatus for obtaining video data, which is not described in detail herein.
While particular embodiments are described above, it will be understood it is not intended to limit the present application to these particular embodiments. On the contrary, the present application includes alternatives, modifications and equivalents that are within the spirit and scope of the appended claims. Numerous specific details are set forth in order to provide a thorough understanding of the subject matter presented herein. But it will be apparent to one of ordinary skill in the art that the subject matter may be practiced without these specific details. In other instances, well-known methods, procedures, components, and circuits have not been described in detail so as not to unnecessarily obscure aspects of the embodiments.
The terminology used in the description of the present application herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the present application. As used in the description of the present application and the appended claims, the singular forms "a, " "an, " and "the" are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will also be understood that the term "and/or" as used herein refers to and encompasses any and all possible combinations of one or more of the associated listed items. It will be further understood that the terms "includes, " "including, " "comprises, " and/or "comprising, " when used in this specification, specify the presence of stated features, operations, elements, and/or components, but do not  preclude the presence or addition of one or more other features, operations, elements, components, and/or groups thereof.
As used herein, the term "if" may be construed to mean "when" or "upon" or "in response to determining" or "in accordance with a determination" or "in response to detecting, " that a stated condition precedent is true, depending on the context. Similarly, the phrase "if it is determined [that a stated condition precedent is true] " or "if [a stated condition precedent is true] " or "when [a stated condition precedent is true] "may be construed to mean "upon determining" or "in response to determining" or "in accordance with a determination" or "upon detecting" or "in response to detecting" that the stated condition precedent is true, depending on the context.
Although some of the various drawings illustrate a number of logical stages in a particular order, stages that are not order dependent may be reordered and other stages may be combined or broken out. While some reordering or other groupings are specifically mentioned, others will be obvious to those of ordinary skill in the art and so do not present an exhaustive list of alternatives. Moreover, it should be recognized that the stages could be implemented in hardware, firmware, software or any combination thereof.
The foregoing description, for purpose of explanation, has been described with reference to specific embodiments. However, the illustrative discussions above are not intended to be exhaustive or to limit the present application to the precise forms disclosed. Many modifications and variations are possible in view of the above teachings. The embodiments were chosen and described in order to best explain the principles of the present application and its practical applications, to thereby enable others skilled in the art to best utilize the present application and various embodiments with various modifications as are suited to the particular use contemplated. 

Claims (15)

  1. A method of obtaining and sharing video data using a social networking application comprising:
    at a portable computer device having one or more processors and memory storing programs executed by the one or more processors,
    recording a plurality of video segments successively;
    marking each video segment with a respective identifier;
    caching the plurality of video segments in the memory for processing;
    determining whether a total time length of the plurality of video segments exceeds a pre-set threshold;
    stopping video recording if the total time length of the plurality of video segments exceeds the pre-set threshold; and
    integrating the plurality of video segments into one video according to a predefined order.
  2. The method of obtaining and sharing video data using a social networking application of claim 1, wherein determining whether the total time length of the plurality of video segments exceeds the pre-set threshold further comprises:
    configuring a timer with a value equal to the pre-set threshold;
    decrementing the timer when each video segment is recorded;
    determining whether the value of the timer equals to or less than zero; and
    when the value of the timer equals to or less than zero, indicating that the total length of time of the plurality of video segments exceeds the pre-set threshold.
  3. The method of obtaining and sharing video data using a social networking application of claim 1, wherein integrating the plurality of video segments into the one video according to a predefined order further comprises:
    reading the plurality of identifiers that correspond to the plurality of video segments, respectively;
    re-organizing the plurality of video segments into a sequence based on sorting of the plurality of identifiers; and
    aggregating the re-organized plurality of video segments into the one video. 
  4. The method of obtaining and sharing video data using a social networking application of claim 1, wherein integrating the plurality of video segments into the one video according to a predefined order further comprises:
    reading a plurality of recording times that correspond to the plurality of video segments, respectively;
    re-organizing the plurality of video segments into a sequence based on sorting of the plurality of recording times; and
    aggregating the re-organized plurality of video segments into the one video.
  5. The method of obtaining and sharing video data using a social networking application of claim 1, further comprises:
    prompting a message to a user of the social networking application whether to share the one video in a social network; and
    when an instruction from the user to share the one video in the social network is received, choosing a participant of the social network that the user wants to share with, and sending the one video to the participant via the social network application.
  6. A system of obtaining and sharing video data using a social networking application comprising:
    one or more processors;
    memory; and
    one or more programs stored in the memory and configured for execution by the one or more processors, the one or more programs including instructions for:
    recording a plurality of video segments successively;
    marking each video segment with a respective identifier;
    caching the plurality of video segments in the memory for processing;
    determining whether a total time length of the plurality of video segments exceeds a pre-set threshold;
    stopping video recording if the total time length of the plurality of video segments exceeds the pre-set threshold; and
    integrating the plurality of video segments into one video according to a predefined order.
  7. The system of obtaining and sharing video data using a social networking application of claim 6, wherein the one or more programs further include instructions for:
    configuring a timer with a value equal to the pre-set threshold;
    decrementing the timer when each video segment is recorded;
    determining whether the value of the timer equals to or less than zero; and
    when the value of the timer equals to or less than zero, indicating that the total time length of the plurality of video segments exceeds the pre-set threshold.
  8. The system of obtaining and sharing video data using a social networking application of claim 6, wherein the one or more programs further include instructions for:
    reading the plurality of identifiers that correspond to the plurality of video segments, respectively;
    re-organizing the plurality of video segments into a sequence based on sorting of the plurality of identifiers; and
    aggregating the re-organized plurality of video segments into the one video.
  9. The system of obtaining and sharing video data using a social networking application of claim 6, wherein the one or more programs further include instructions for:
    reading a plurality of recording times that correspond to the plurality of video segments, respectively;
    re-organizing the plurality of video segments into a sequence based on sorting of the plurality of recording times; and
    aggregating the re-organized plurality of video segments into the one video.
  10. The system of obtaining and sharing video data using a social networking application of claim 6, wherein the one or more programs further include instructions for:
    prompting a message to a user of the social networking application whether to share the one video in a social network; and
    when an instruction from the user to share the one video in the social network is received, choosing a participant of the social network that the user wants to share with, and sending the one video to the participant via the social network application.
  11. A non-transitory computer readable storage medium, storing one or more programs for execution by one or more processors of a system of obtaining and sharing video data using a social networking application, the one or more programs including instructions for:
    recording a plurality of video segments successively;
    marking each video segment with a respective identifier;
    caching the plurality of video segments in the memory for processing;
    determining whether a total time length of the plurality of video segments exceeds a pre-set threshold;
    stopping video recording if the total time length of the plurality of video segments exceeds the pre-set threshold; and
    integrating the plurality of video segments into one video according to a predefined order.
  12. The non-transitory computer readable storage medium of claim 11, wherein the one or more programs including instructions for:
    configuring a timer with a value equal to the pre-set threshold;
    decrementing the timer when each video segment is recorded;
    determining whether the value of the timer equals to or less than zero; and
    when the value of the timer equals to or less than zero, indicating that the total time length of the plurality of video segments exceeds the pre-set threshold.
  13. The non-transitory computer readable storage medium of claim 11, wherein the one or more programs including instructions for:
    reading the plurality of identifiers that correspond to the plurality of video segments, respectively;
    re-organizing the plurality of video segments into a sequence based on sorting of the plurality of identifiers; and
    aggregating the re-organized plurality of video segments into the one video.
  14. The non-transitory computer readable storage medium of claim 11, wherein the one or more programs including instructions for:
    reading a plurality of recording times that correspond to the plurality of video segments, respectively; 
    re-organizing the plurality of video segments into a sequence based on sorting of the plurality of recording times; and
    aggregating the re-organized plurality of video segments into the one video.
  15. The non-transitory computer readable storage medium of claim 11, wherein the one or more programs including instructions for:
    prompting a message to a user of the social networking application whether to share the one video in a social network; and
    when an instruction from the user to share the one video in the social network is received, choosing a participant of the social network that the user wants to share with, and sending the one video to the participant via the social network application. 
PCT/CN2014/086719 2013-09-27 2014-09-17 Method, apparatus, and terminal for obtaining video data WO2015043403A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201310452594.5 2013-09-27
CN201310452594.5A CN104519262B (en) 2013-09-27 2013-09-27 Obtain the method, apparatus and terminal of video data

Publications (1)

Publication Number Publication Date
WO2015043403A1 true WO2015043403A1 (en) 2015-04-02

Family

ID=52742032

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2014/086719 WO2015043403A1 (en) 2013-09-27 2014-09-17 Method, apparatus, and terminal for obtaining video data

Country Status (2)

Country Link
CN (1) CN104519262B (en)
WO (1) WO2015043403A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105979383A (en) * 2016-06-03 2016-09-28 北京小米移动软件有限公司 Image acquisition method and device
CN111857914A (en) * 2020-07-06 2020-10-30 上海意略明数字科技股份有限公司 Method and device for acquiring operation information of application program on terminal, storage medium and terminal
JP2022508876A (en) * 2018-11-08 2022-01-19 北京微播視界科技有限公司 Video playback speed control methods, devices, devices and storage media

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105187692B (en) * 2014-06-16 2018-05-18 腾讯科技(北京)有限公司 Video capture method and device
CN104980794A (en) * 2015-06-30 2015-10-14 北京金山安全软件有限公司 Video splicing method and device
CN106131432A (en) * 2016-08-10 2016-11-16 深圳市金立通信设备有限公司 A kind of video capture method and terminal
CN108322659B (en) * 2018-03-30 2021-05-28 维沃移动通信有限公司 Shooting method and mobile terminal
CN108616696B (en) * 2018-07-19 2020-04-14 北京微播视界科技有限公司 Video shooting method and device, terminal equipment and storage medium
CN108965706B (en) * 2018-07-19 2020-07-07 北京微播视界科技有限公司 Video shooting method and device, terminal equipment and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101175214A (en) * 2007-11-15 2008-05-07 北京大学 Method and apparatus for real-time detecting advertisement from broadcast data stream
CN101217638A (en) * 2007-12-28 2008-07-09 深圳市迅雷网络技术有限公司 A downloading method, system and device of video file fragmentation
CN101316334A (en) * 2008-07-07 2008-12-03 中兴通讯股份有限公司 Method for processing television media data and television mobile phone
CN102196008A (en) * 2010-03-08 2011-09-21 株式会社日立制作所 Peer-to-peer downloading method, video equipment and content transmission method

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101351091B1 (en) * 2006-12-22 2014-01-14 삼성전자주식회사 Image forming apparatus and control method of consecutive photographing image
JP4336387B2 (en) * 2007-01-15 2009-09-30 パナソニック株式会社 Imaging device
JP4998026B2 (en) * 2007-03-15 2012-08-15 ソニー株式会社 Image processing apparatus, imaging apparatus, image display control method, and computer program
US8072487B2 (en) * 2007-04-25 2011-12-06 Fujifilm Corporation Picture processing apparatus, picture recording apparatus, method and program thereof
FR2918240A1 (en) * 2007-06-26 2009-01-02 Thomson Licensing Sa METHOD FOR CREATING A SOUND SUITE OF PHOTOGRAPHS, AND APPARATUS FOR CREATING SUCH A SOUND SUITE
US9224425B2 (en) * 2008-12-17 2015-12-29 Skyhawke Technologies, Llc Time stamped imagery assembly for course performance video replay
JP5508128B2 (en) * 2010-05-10 2014-05-28 キヤノン株式会社 Video recording apparatus and control method thereof
JP5054175B2 (en) * 2010-09-08 2012-10-24 キヤノン株式会社 Imaging apparatus and control method thereof, imaging control apparatus, and imaging control method
CN103310005A (en) * 2013-06-28 2013-09-18 北京小米科技有限责任公司 Thumbnail display method and device

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101175214A (en) * 2007-11-15 2008-05-07 北京大学 Method and apparatus for real-time detecting advertisement from broadcast data stream
CN101217638A (en) * 2007-12-28 2008-07-09 深圳市迅雷网络技术有限公司 A downloading method, system and device of video file fragmentation
CN101316334A (en) * 2008-07-07 2008-12-03 中兴通讯股份有限公司 Method for processing television media data and television mobile phone
CN102196008A (en) * 2010-03-08 2011-09-21 株式会社日立制作所 Peer-to-peer downloading method, video equipment and content transmission method

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105979383A (en) * 2016-06-03 2016-09-28 北京小米移动软件有限公司 Image acquisition method and device
CN105979383B (en) * 2016-06-03 2019-04-30 北京小米移动软件有限公司 Image acquiring method and device
JP2022508876A (en) * 2018-11-08 2022-01-19 北京微播視界科技有限公司 Video playback speed control methods, devices, devices and storage media
EP3879838A4 (en) * 2018-11-08 2022-08-03 Beijing Microlive Vision Technology Co., Ltd Video play speed control method, apparatus and device, and storage medium
JP7169456B2 (en) 2018-11-08 2022-11-10 北京微播視界科技有限公司 Video playback speed control method, device, device and storage medium
CN111857914A (en) * 2020-07-06 2020-10-30 上海意略明数字科技股份有限公司 Method and device for acquiring operation information of application program on terminal, storage medium and terminal
CN111857914B (en) * 2020-07-06 2021-10-01 上海意略明数字科技股份有限公司 Method and device for acquiring operation information of application program on terminal, storage medium and terminal

Also Published As

Publication number Publication date
CN104519262B (en) 2018-04-27
CN104519262A (en) 2015-04-15

Similar Documents

Publication Publication Date Title
US10419823B2 (en) Method for controlling multimedia playing, apparatus thereof and storage medium
US10635449B2 (en) Method and apparatus for running game client
WO2015043403A1 (en) Method, apparatus, and terminal for obtaining video data
US10193834B2 (en) Method and apparatus for downloading and displaying pictures
CN105786878B (en) Display method and device of browsing object
CN104965641B (en) information display method and device
CN106059894B (en) Message processing method and device
US9794206B2 (en) Method and apparatus for managing the display of messages of a group chat
CN104967896A (en) Method for displaying bulletscreen comment information, and apparatus thereof
US20140365892A1 (en) Method, apparatus and computer readable storage medium for displaying video preview picture
CN107193664B (en) Message display method and device and mobile terminal
CN106375179B (en) Method and device for displaying instant communication message
US20150019764A1 (en) Information displaying method, mobile terminal device and non-transitory computer readable storage medium
US20170364257A1 (en) Gesture control method, gesture control device and gesture control system
CN106708554B (en) Program operating method and device
JP2018500644A (en) Method, apparatus, and terminal for matching facial expression images
WO2014194713A1 (en) Method,apparatus and computer readable storage medium for displaying video preview picture
TW201504918A (en) Method, device and moving terminal for examining messages
CN104066009A (en) Method, device, terminal, server and system for program identification
US20170171255A1 (en) Method and device for establishing a data connection and computer-readable medium
CN109714476B (en) Data processing method and device, mobile terminal and storage medium
CN103399705A (en) Method, device and equipment for remotely controlling terminal equipment
CN105577712B (en) A kind of file uploading method, device and system
CN106294528B (en) Method and device for realizing information transmission
CN111966271B (en) Screen panorama screenshot method and device, terminal equipment and storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14849075

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205N DATED 19/08/2016)

122 Ep: pct application non-entry in european phase

Ref document number: 14849075

Country of ref document: EP

Kind code of ref document: A1