WO2022088447A1 - 视频播放方法、系统、电子设备和存储介质 - Google Patents

视频播放方法、系统、电子设备和存储介质 Download PDF

Info

Publication number
WO2022088447A1
WO2022088447A1 PCT/CN2020/138146 CN2020138146W WO2022088447A1 WO 2022088447 A1 WO2022088447 A1 WO 2022088447A1 CN 2020138146 W CN2020138146 W CN 2020138146W WO 2022088447 A1 WO2022088447 A1 WO 2022088447A1
Authority
WO
WIPO (PCT)
Prior art keywords
video
picture
image
notification
original video
Prior art date
Application number
PCT/CN2020/138146
Other languages
English (en)
French (fr)
Inventor
吴文宪
Original Assignee
广州视源电子科技股份有限公司
广州视睿电子科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 广州视源电子科技股份有限公司, 广州视睿电子科技有限公司 filed Critical 广州视源电子科技股份有限公司
Publication of WO2022088447A1 publication Critical patent/WO2022088447A1/zh

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/15Conference systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content

Definitions

  • the embodiments of the present disclosure relate to the technical field of video processing, and in particular, to a video playback method, system, electronic device, and storage medium.
  • video conferencing can meet the needs of people to communicate anytime and anywhere.
  • a large amount of information such as videos and images usually needs to be transmitted, and information transmission between multiple users is usually required.
  • video screen playback process in the video conference requires more network transmission resources, which is easy to cause network congestion, resulting in slow video screen playback and freezes.
  • a common improvement measure is that the video picture sender increases the detection of the region of interest in the video picture to be sent, and then maintains the picture quality of the region of interest and reduces the picture quality of the non-interested region, thereby reducing the video bit rate to improve the occupation of network transmission resources.
  • this solution can only slightly improve the occupancy of network transmission resources during the video conference, and cannot improve the utilization of network transmission resources.
  • a video playback method comprising:
  • another video playback method comprising:
  • the second original video image is downgraded to obtain a downgraded video image, wherein the unwatched notification is that the first device is playing the
  • the first original video image is collected, local sensing data is collected, and it is determined according to the local sensing data that when there is no one at the first device, a notification is generated, wherein the image degradation includes reducing the resolution and/or or reduce the frame rate;
  • a video playback system in a third aspect, includes a first device and a second device;
  • the second device configured to send the first original video picture to the first device
  • the first device is configured to collect local sensing data when the first original video image is played, and determine whether there is no one at the first device according to the local sensing data, and when it is determined that the first device is unoccupied There is no one at the device, generate a no-view notification, and send the no-view notification to the second device;
  • the second device is further configured to downgrade the second original video image to obtain a downgraded video image when receiving the unwatched notification, and return the downgraded video to the first device A picture, wherein the picture degradation includes reducing the resolution and/or reducing the frame rate;
  • the first device is further configured to play the degraded video picture.
  • a video playback device comprising:
  • a collection module configured to collect local sensing data when the first device plays the first original video image
  • a determining module configured to determine whether there is no one at the first device according to the local sensing data
  • a notification generation module configured to generate an unmanned viewing notification when it is determined that there is no one at the first device
  • a notification sending module configured to send the unwatched notification to the second device, so that when the second device receives the unwatched notification, the second original video image is degraded to obtain the degraded quality a video picture, and returning the downgraded video picture to the first device, wherein the downgrading of the picture includes reducing the resolution and/or reducing the frame rate;
  • a playing module configured to play the degraded video picture.
  • another video playback device is also provided, and the device includes:
  • a first sending module configured to send the first original video picture to the first device
  • the picture degradation module is configured to perform picture degradation on the second original video picture when receiving the unwatched notification sent by the first device to obtain a downgraded video picture, wherein the unwatched notification is all
  • the first device collects local sensing data when playing the first original video image, and determines according to the local sensing data that a notification is generated when there is no one at the first device, wherein the image is down quality including reduced resolution and/or reduced frame rate;
  • the second sending module is configured to send the degraded video picture to the first device.
  • an electronic device comprising: a memory and one or more processors;
  • the memory configured to store one or more programs
  • the one or more processors When the one or more programs are executed by the one or more processors, the one or more processors perform the following operations:
  • a non-volatile storage medium containing computer-executable instructions is provided, a computer program is stored thereon, and when the computer program is executed by a processor, the following steps are implemented:
  • the above video playback method, system, electronic device and storage medium can detect the first device by collecting local sensing data when the first device plays the first original video image, and determining whether there is no one at the first device according to the local sensing data. Whether there is no one watching on the device side, when it is determined that no one is watching the first device, generate a no one watching notification, send the no one watching notification to the second device, and notify the second device when no one is watching on the first device side. When receiving the notification, the device downgrades the second original video image to obtain the downgraded video image, and returns the downgraded video image to the first device. When people watch, the second device degrades the video quality of the subsequently played video, reducing the video data that needs to be transmitted, so as to better improve the occupation of network transmission resources and improve the utilization rate of network transmission resources during the video conference. .
  • the above-mentioned video playback method does not need to perform too much analysis and processing on the video picture, and can save data. Process resources.
  • the above-mentioned video playback method can degrade the picture when the user temporarily leaves, reduce the data transmission bandwidth, and restore the picture quality when the user is detected to return, thereby reducing the user's manual operation on picture quality and improving the user experience.
  • the bandwidth of data transmission is reduced, and the utilization rate of network transmission resources is improved.
  • FIG. 1 is a schematic diagram of an application environment of a video playback method provided in Embodiment 1 of the present disclosure
  • FIG. 2 is a flowchart of a video playback method provided in Embodiment 1 of the present disclosure
  • FIG. 3 is a schematic diagram of an application environment of a video playback method provided in Embodiment 1 of the present disclosure
  • FIG. 4 is a schematic diagram of a scenario in which a user leaves the field during video playback according to Embodiment 1 of the present disclosure
  • Embodiment 1 of the present disclosure is a schematic diagram of a scenario of multi-user video playback provided by Embodiment 1 of the present disclosure
  • FIG. 6 is a schematic diagram of an application environment of a video playback method provided in Embodiment 2 of the present disclosure
  • FIG. 7 is a flowchart of a video playback method provided in Embodiment 3 of the present disclosure.
  • FIG. 8 is a schematic diagram of a scene of a user returning to the scene when a video is played according to Embodiment 3 of the present disclosure
  • FIG. 9 is a sequence diagram of a video playback method provided in Embodiment 3 of the present disclosure.
  • FIG. 10 is a sequence diagram of a server-forwarded video playback method provided in Embodiment 3 of the present disclosure.
  • FIG. 11 is a schematic diagram of a video playback method provided in Embodiment 3 of the present disclosure.
  • FIG. 12 is a flowchart of a video playback method provided in Embodiment 4 of the present disclosure.
  • FIG. 13 is a schematic structural diagram of a video playback device according to Embodiment 6 of the present disclosure.
  • FIG. 14 is a schematic structural diagram of a video playback device according to Embodiment 7 of the present disclosure.
  • FIG. 15 is a schematic structural diagram of an electronic device according to Embodiment 8 of the present disclosure.
  • FIG. 16 is a schematic structural diagram of an electronic device according to Embodiment 9 of the present disclosure.
  • FIG. 1 is a schematic diagram of an application environment of a video playback method provided by Embodiment 1 of the present disclosure.
  • the local device 102 and the remote device 104 may include, but are not limited to, various computers, mobile phones, tablets equipped with cameras Or projector, the local device 102 and the remote device 104 communicate through wired or wireless networks, either directly or through a server, the display screen of the local device 102 can display the video images of users B and C in the remote location, and the remote device The display screen of 104 can display at least the video picture of the local user A, the local device 102 can also display the videos, pictures and documents shared by the remote device 104, and the remote device 104 can also display the videos, pictures and documents shared by the local device 102. .
  • the following describes the video playing process by taking the local device 102 playing the video picture of user B as an example for the video conference scenario.
  • the remote device 104 can use the camera to shoot the user B in real time, obtain the video of the user B, and record the video of the user B according to the set resolution. and the frame rate are adjusted to generate a video picture of user B with a fixed resolution and a fixed frame rate.
  • the remote device 104 can use a video encoder to encode the video picture of user B, and the video encoder compresses the video picture according to the preset bit rate, and generates a data volume.
  • the offsite device 104 may transmit the video encoded data to the local device 102 .
  • the local device 102 can decode it through a video decoder to restore the video image of user B for playback by the local device 102, and the video image of user B can be displayed on the local device 102 in real time.
  • the video shooting process for user B can be omitted, and the video encoder can directly encode the picture displayed on the remote device 104 and transmit the encoded data.
  • the local device 102 After decoding it, the local device 102 can display the video, picture or document shared by the remote device 104 in real time.
  • the process of the remote device 104 playing the video screen of the user A and sharing the video, picture or document of the local device may be the reverse process of the above process, which will not be repeated here.
  • the remote device 104 may generate a video picture with a higher resolution and/or frame rate, and transmit it to the local device 102 for playback.
  • the video encoding bit rate remains unchanged, the amount of data that needs to be transmitted during the video playback process is large, which occupies more network transmission resources.
  • the area of interest is usually detected in the video to be sent, the image quality of the area of interest is maintained, and the image quality of the non-interested area is reduced to improve the occupation of network transmission resources, but this method still cannot improve the Regarding the occupation of network transmission resources, it is difficult to improve the utilization rate of network transmission resources.
  • FIG. 2 is a flowchart of a video playback method provided by Embodiment 1 of the present disclosure.
  • FIG. 3 is a schematic diagram of an application environment of a video playback method provided by Embodiment 1 of the present disclosure.
  • a user conducts a video conference through at least one first device 302 and at least one second device 304, wherein the first device 302 may be a local terminal device and the second device 304 may be a remote terminal device.
  • the video playback method provided in this embodiment may be executed by a terminal device, and the terminal device may be implemented by software and/or hardware, wherein the software may be implemented by running software on hardware, and the terminal device may be composed of two It is composed of multiple physical entities, or it can be composed of one physical entity.
  • the terminal device may be a computer, a mobile phone, a tablet, or a projector.
  • the video playback method in Embodiment 1 of the present disclosure taking the application to the first device 302 in FIG. 3 as an example, may include the following steps:
  • Step S210 when the first device plays the first original video picture, collect local sensing data.
  • the original video picture may be a video picture with high display quality
  • the first original video picture may be an original video picture collected by the second device at the current moment.
  • the local sensing data may be data collected by a first device, wherein the first device may collect local sensing data through at least one of an image collection device, an audio collection device, and an infrared sensor device.
  • the second device can collect videos of users in different places to obtain original video images with high display quality.
  • the original video images collected at the current moment can be determined. It is the first original video picture, and after a certain collection period reaches the next moment of the current moment, the original video picture collected at the next moment can be determined as the second original video picture, and so on.
  • the second device may send the first original video picture with high display quality to the first device, and after receiving the first original video picture, the first device plays the first original video picture, and the first original video picture displayed on the screen of the first device.
  • the first device can be configured with at least one of an image capture device, an audio capture device, and an infrared sensing device.
  • the first device can collect local image frame data through the image capture device , the image frame data is used as local sensing data, the audio data around the first device can also be collected through an audio collection device, the collected audio data can be used as local sensing data, and the surrounding area of the first device can also be collected through infrared sensors. The collected infrared data is used as the local sensing data.
  • the first device may be configured with a front-facing camera.
  • the front-facing camera may capture a video within a certain range in front of the display screen of the first device to obtain image frame data.
  • the above-mentioned original video pictures may not be limited to the collected video pictures of remote users, and the original video pictures may also be videos shared by the second device, Pictures and pictures of documents.
  • Step S220 according to the local sensing data, determine whether there is no one at the first device.
  • Step S230 when it is determined that there is no one at the first device, a notification of no one watching is generated.
  • the first device may determine whether there is a person or no one at the first device according to local sensing data, and generate a no-watch notification when no one is there.
  • the first device may collect a local video through an image acquisition device, and intercept one frame of image or multiple consecutive frames of images from the local video.
  • image frame data the first device may also use an image acquisition device Directly collect local image frame data.
  • the first device can detect whether there is a face in the image frame data, and if a human face is detected, it can be determined that there is a person in the first device, and if no human face is detected, it can be determined that the first device is located Unattended, and generate an unattended notification on the first device. For example, when the first device plays the first original video image, it can capture consecutive N frames of images in the local video. If a face is detected in any frame, it can be determined that there is someone at the first device. If no face is detected in the image, it can be determined that there is no one at the first device, and a notification of no one watching is generated.
  • the first device can count the duration of unspeaking (silence duration) in the audio data, and compare the duration with the preset duration threshold. In contrast, if the time length does not exceed the time length threshold, it can be determined that there is someone at the first device, and if the time length exceeds the time length threshold, it can be determined that there is no one at the first device, and a no-view notification is generated on the first device.
  • the first device can compare the collected audio data with a preset audio range, and can determine that no one speaks if the collected audio data is not within the preset audio range.
  • the first device when the first device plays the first original video picture, it can start to collect audio around the first device at intervals of 1 second, and identify whether there is a human voice in the audio, if there is no voice, start timing and count the no voice If the length of time does not exceed the time length threshold of 10 seconds, it can be determined that there is someone at the first device; otherwise, if the length of time exceeds 10 seconds, it can be determined that there is no one at the first device, and the first device generates no one. Watch notifications. It is also possible to set an audio range of 500-2000 Hz (Hertz) in the first device, and compare the collected audio data with it. If the audio data is 400 Hz, which is not within the above audio range, it can be determined that no one speaks.
  • 500-2000 Hz Hertz
  • the first device can compare the infrared data with a preset infrared data interval, and if the infrared data is within the infrared data interval, it can determine There is someone at the first device, and if the infrared data is outside the infrared data interval, it can be determined that there is no one at the first device, and a notification of no viewing is generated at the first device.
  • the temperature around the first device can be measured by an infrared thermometer, and if the measured temperature is within the human body temperature range, it can be determined that there is a person at the first device, otherwise, If the measured temperature is not within the temperature range of the human body, it may be determined that there is no one at the first device, and a notification of no one watching is generated.
  • Step S240 sending a no-viewing notification to the second device, so that when the second device receives the no-viewing notification, it degrades the second original video image to obtain a downgraded video image, and returns it to the first device Downgrading the video picture, wherein the picture degradation includes reducing the resolution and/or reducing the frame rate.
  • Step S250 playing the degraded video picture.
  • the picture degradation may be an operation of reducing the picture display quality, including reducing the resolution and/or reducing the frame rate.
  • the downgraded video picture may display a video picture with reduced quality for the picture.
  • the second original video picture may be an original video picture collected by the second device at the next moment of the current moment.
  • the second device may be a terminal device or a server device.
  • the first device may send the unwatched notification to the second device, and when receiving the unwatched notification, the second device may update the captured second original video image at the next moment of the current moment.
  • the second device can send the degraded video picture to the first device, and the first device can play the degraded video picture when receiving the degraded video picture sent by the second device to display the degraded video image on the screen of the first device.
  • the second device may reduce the resolution of the high-quality video images collected at the next moment when receiving a notification that no one is watching, generate a reduced-quality video image with a lower resolution, and send it to the first device for playback;
  • the second device can also reduce the frame rate of the high-quality video images collected at the next moment to generate a lower-quality video image with a lower frame rate, and send them to the first device for playback.
  • the downgraded video picture may not be transmitted to the first device, in which case the first device displays a still video. screen.
  • the first device detects that the first original video image is watched by someone, there is no need to send a notification, and the second device can directly send the captured second original video image to the first device at the next moment.
  • the second original video picture can be played. If the first device detects that the first original video image has changed from unwatched to someone watched, it can also send a watched notification to the second device, and the second device can stop the image degradation at the next moment when it receives the watched notification. , and directly send the captured second original video image to the first device for the first device to play.
  • FIG. 4 is a schematic diagram of a scenario in which a user leaves the field during video playback according to Embodiment 1 of the present disclosure.
  • the resolution of the original video screen played by the first device 302 at the current moment can be set to 4000 pixels, and the frame rate is 25 frames per second (Frame Per Second, referred to as FPS for short).
  • FPS Frame Per Second
  • the first device 302 In real-time detection, whether there is human body information in the first device 302 can be detected in real time.
  • the first device 302 can collect local video, and determine whether there is human body information in the first device 302 by detecting whether there is human body movement, human body shape, or face image in the local video.
  • human body information exists in the first device 302 can also be determined by whether human body infrared data or human voice can be detected within a certain range around the first device 302 . If the first device 302 does not detect human body information, it can be determined that there is no one in the first device 302 at the current moment, and no one is watching the original video image.
  • the first device 302 can also be configured with a touch screen, by setting a time threshold, when the touch screen does not receive the user's touch screen operation and the duration exceeds the time threshold, it can be determined that the original video picture at the current moment is unwatched, otherwise It can be judged that someone is watching. As shown in FIG.
  • the second device 304 when the local user A leaves and the first device 302 detects that the original video picture at the current moment is unwatched, the second device 304 can be notified, and the second device 304 can display the notification on the display screen when receiving the notification Display that local user A has left, and downgrade the original video at the next moment. For example, you can reduce the resolution to 2000 pixels, or reduce the frame rate to 15FPS, and send the downgraded video to the first The device 302 performs display.
  • the first device 302 can display a low-resolution image of the remote user B, and the second device 304 can also reduce the frame rate to 0, and do not transmit video images to the first device 302. In this case, the first A device 302 displays still video footage.
  • FIG. 5 is a schematic diagram of a scenario of multi-user video playback provided by Embodiment 1 of the present disclosure.
  • this scenario includes a first device 302 and at least two second devices 304.
  • the first device 302 has a local user A
  • the second device 304 has remote users B and C.
  • the first device 302 can send a no-view notification to all the second devices 304, and when the second device 304 receives the no-view notification, it can display on the display that the local user A has left, and normally display the user B who has not left. or C.
  • the second device 304 may further downgrade the respective original video pictures according to different resolutions/frame rates, generate downgraded video pictures, and send the downgraded video pictures to the first device 302 respectively, where the first device 302
  • the degraded video images of remote users B and C are displayed on the .
  • the first device plays the first original video image
  • local sensing data is collected, and whether there is no one on the first device can be detected according to the local sensing data.
  • Watch when it is determined that there is no one at the first device, generate a no-view notification, send a no-view notification to the second device, and notify the second device when no one is watching on the first device, and the second device receives the notification
  • the second original video image is degraded
  • the degraded video image is obtained, and the degraded video image is returned to the first device, and the first device plays the degraded video image.
  • the second device degrades the image of the subsequently played video to reduce the video data to be transmitted, so as to better improve the occupation of network transmission resources and improve the utilization rate of network transmission resources during the video conference.
  • the above-mentioned video playback method does not need to perform too much analysis and processing on the video picture, and can save data. Process resources.
  • the above-mentioned video playback method can degrade the picture when the user temporarily leaves, reduce the data transmission bandwidth, and restore the picture quality when the user is detected to return, thereby reducing the user's manual operation on picture quality and improving the user experience.
  • the bandwidth of data transmission is reduced, and the utilization rate of network transmission resources is improved.
  • the local sensing data includes image frame data collected by an image collection device of the first device.
  • the image capturing device may be a device that captures video or images around the first device, and may include, but is not limited to, various cameras and video cameras.
  • the first device may be configured with an image capture device.
  • the first device may capture a local video through the image capture device, and intercept one frame of image or multiple consecutive images from the local video.
  • Frame images, as image frame data the first device may also directly collect local image frame data through an image acquisition device.
  • the first device may be configured with a front-facing camera.
  • the front-facing camera captures a video within a certain range in front of the display screen of the first device, and the image frame data is obtained by intercepting image frames in the video. It is also possible to directly obtain image frame data by shooting an image within a certain range in front of the display screen of the first device through the front camera.
  • the local sensing data includes image frame data collected by the image acquisition device of the first device, which can facilitate the detection of local unmanned conditions.
  • step S220 includes:
  • the face detection may be to detect whether there is a face image in the image.
  • the first device can sequentially obtain image frame data from the local video, and detect whether there is a face in the image frame data. If a face image is detected, it can be determined that there is a person in the first device, and if no human is detected face image, it can be determined that there is no one at the first device.
  • N frames of images can be captured from the local video. If a face is detected in any frame, it can be determined that there is someone at the first device. If no face image is detected, it can be determined that there is no one at the first device. In order to improve the detection accuracy, when a face image is detected in at least M (M ⁇ N) frames in the N frame images, it is determined that there is a person at the first device, and if a face image is detected in less than M frames, Then it can be determined that there is no one at the first device.
  • the face detection can also be used to detect whether the face in the image is a specified face image. For example, it can detect whether the face image in the image frame data matches the face image of the local user A. If If there is a match, it can be determined that there is someone at the first device, and if it does not match, it is determined that there is no one at the first device.
  • the technical solutions of the embodiments of the present disclosure by sequentially acquiring image frame data and performing face detection on the image frame data, it can be detected whether there is a face image in the image frame data, and when a face image is detected in the image frame data , determine that there is a person at the first device, and when no face image is detected in the acquired image frame data for a preset number of consecutive times, it is determined that there is no one at the first device, and the first device can be determined according to the result of the face detection. Whether there is no one, in order to reduce the quality of the displayed picture and reduce the video data to be transmitted when there is no one, so as to better improve the occupation of network transmission resources and improve the utilization rate of network transmission resources during the video conference process.
  • the local sensing data further includes audio data collected by the audio collection device of the first device
  • the step S220 further includes:
  • Count the silence duration in the audio data where the silence duration is the duration when no sound is detected in the audio data; compare the silence duration with a preset duration threshold; when the silence duration does not exceed the duration threshold, determine that the first device There is someone; when the silence duration exceeds the duration threshold, it is determined that there is no one at the first device.
  • the audio collection device may be a device that collects audio around the first device, which may include, but is not limited to, various microphones.
  • the audio data may be sounds collected around the first device.
  • the silent duration may be the duration of no one speaking in the audio data.
  • the first device may be configured with an audio collection device.
  • the first device collects local audio data through the audio collection device, and uses the audio data as local sensor data.
  • the first device can also compare the audio data with the audio range of the human voice. If the local audio is not within the audio range of the human voice, it can be determined that no one speaks, and the first device can count the length of time that no one speaks in the audio data. , get the silence duration, compare the silence duration with the preset duration threshold, if the silence duration does not exceed the duration threshold, it can be determined that there is someone at the first device, and if the silence duration exceeds the duration threshold, it can be determined that there is no one at the first device .
  • the first device may be configured with a microphone, and when the first device plays the first original video image, the microphone collects audio data within a certain range around the first device, and the audio range can also be set to 500-2000 Hz, and the collected audio The data is compared with the audio frequency range. If the audio data is 400 Hz, which is not within the above audio frequency range, it can be determined to be silent.
  • the first device plays the first original video image, it can start to collect audio data around the first device at intervals of 1 second, and identify whether the audio data is silent. If it is silent, start timing and count the silent duration. If the duration does not exceed the duration threshold of 10 seconds, it can be determined that there is someone at the first device; otherwise, if the silent duration exceeds 10 seconds, it can be determined that there is no one at the first device.
  • the audio data is the audio data of the specified user. For example, it can be detected whether the audio data matches the audio data of the local user A. If it matches, it can be determined that the local user A is not silent. , it can be determined that the local user A is silent, and by counting the silent duration of the local user A, it is determined whether there is no one at the first device.
  • the silence duration in the audio data is counted, and the silence duration is compared with a preset duration threshold.
  • the silence duration does not exceed the duration threshold, it is determined that there is someone at the first device.
  • the time duration threshold is determined, it is determined that there is no one at the first device, and it can be determined whether there is no one at the first device according to the statistical result of the silent duration, so as to reduce the display picture quality and reduce the video data that needs to be transmitted when there is no one.
  • the occupancy of network transmission resources is better improved, and the utilization rate of network transmission resources is improved.
  • the local sensing data further includes infrared data collected by an infrared sensing device of the first device
  • the step S220 further includes:
  • the infrared data is within the preset infrared data interval; when the infrared data is within the infrared data interval, it is determined that there is a person at the first device; when the infrared data is not within the infrared data interval, it is determined that there is no one at the first device.
  • the infrared sensing device may be a device that collects infrared data around the first device, which may include, but is not limited to, various infrared thermometers and infrared imagers.
  • the infrared data may be infrared data collected around the first device.
  • the first device may be configured with an infrared sensing device, and when the first device plays the first original video image, the first device collects local infrared data through the infrared sensing device. The first device can also compare the collected infrared data with a preset infrared data interval. If the infrared data is within the infrared data interval, it can be determined that there is someone at the first device, and if the infrared data is outside the infrared data interval, it can be determined. Make sure that no one is at the first device.
  • the first device may be configured with an infrared thermometer.
  • the temperature in a certain range around the first device is measured by the infrared thermometer, which is used as infrared data. If the measured temperature is within the human body temperature range (infrared data range), it can be determined that there is someone at the first device; otherwise, if the measured temperature is not within the human body temperature range, it can be determined that there is no one at the first device.
  • the infrared data is within the preset infrared data range, when the infrared data is within the infrared data range, it is determined that there is a person at the first device, and when the infrared data is not within the infrared data range , it is determined that there is no one at the first device, and whether there is no one at the first device can be determined according to the interval in which the infrared data is located, so as to reduce the quality of the display image and reduce the video data that needs to be transmitted when there is no one, so that during the video conference process
  • the occupancy of network transmission resources is better improved, and the utilization rate of network transmission resources is improved.
  • FIG. 6 is a schematic diagram of an application environment of a video playback method provided by Embodiment 2 of the present disclosure.
  • the user conducts a video conference through the first device 302, the second device 304 and the third device 304, wherein the first device 302 can be a local terminal device, the second device 304 can be a server, and the third device 306 can be For remote terminal equipment.
  • the server may be a server that provides temporary storage and forwarding for video playback, and may be implemented by an independent server or a server cluster composed of multiple servers, or may be a cloud server.
  • the video playback method of Embodiment 2 of the present disclosure may include:
  • the second device is a terminal or a server; when the second device is a terminal, the second original video image is the video image collected by the second device; when the second device is a server, the second original video image is collected by the third device , and send it to the video screen of the second device.
  • the second device 304 may be a remote terminal, and by collecting remote user video, an original video image with high display quality is obtained, including the second original video image. video screen.
  • the third device 306 can be a remote terminal. After collecting the remote user video and obtaining the original video image with high display quality, the third device 306 can send the original video image to The second device 304, the second device 304 stores the original video picture, and forwards it to the first device 302, where the original video picture may include the second original video picture.
  • the remote terminal can send the collected original video picture to the server, and the server stores the original video picture and forwards it to the local terminal.
  • the remote terminal can send the first original video picture to the server at the current moment, and send the second original video picture to the server at the next moment.
  • the server can send the first original video image with high display quality to the local terminal.
  • the local terminal plays the first original video image, and collects local sensor data to obtain the first original video image. Detect whether the first original video picture is watched by anyone, and if it is detected that no one is watching, it can send a no-viewing notification to the server.
  • the original video picture is reduced in resolution and/or downgraded frame rate to obtain a downgraded video picture, and the server can send the downgraded video picture to the local terminal for playback.
  • the local terminal may have local user A
  • the remote terminal may have remote users B and C
  • the local terminal may obtain the video image of the remote terminal from the server.
  • the local terminal may notify the server, and the server receives the notification.
  • can send information to the remote terminal the remote terminal can display on the display screen that the local user A has left, and normally display the user B or C who have not left, and the server can also display the video images of the remote users B and C stored in the server.
  • the quality of the picture is degraded, and the degraded video picture is sent to the local terminal, and the degraded video picture of the remote users B and C can be displayed on the local terminal.
  • the second device is a terminal or a server.
  • the second original video image is a video image collected by the second device.
  • the second original video image is a video image captured by the second device.
  • the original video image is the video image captured by the third device and sent to the second device.
  • the server can perform image degradation of the second original video image, which facilitates centralized management by the server, improves video processing speed, and reduces video playback. delay.
  • FIG. 7 is a flowchart of a video playback method provided in Embodiment 3 of the present disclosure.
  • the video playback method in Embodiment 3 of the present disclosure may include the following steps:
  • Step S710 when the first device plays the first original video picture, collect local sensing data
  • Step S720 according to the local sensing data, determine whether there is no one at the first device
  • Step S730 when it is determined that there is no one at the first device, generate a notification that no one is watching;
  • Step S740 sending a no-viewing notification to the second device, so that when the second device receives the no-viewing notification, the second original video image is degraded to obtain a degraded video image, and returned to the first device Downgrading video images, wherein the image degradation includes lowering resolution and/or lowering frame rate;
  • Step S750 playing the degraded video image
  • Step S760 when the degraded video picture is played on the first device, if it is determined according to the local sensor data that there is a person at the first device, a person watching notification is generated;
  • the first device when the first device is playing a degraded video image, it can collect local sensing data through at least one of an image collection device, an audio collection device, and an infrared sensing device, and by detecting the content of the local sensing data Whether there is human body information including human face, human voice and human body infrared data, it can be determined whether there is someone within a certain range around the first device, and the first device can generate a viewing notification when there is someone. Since the process of determining a person at the first device according to the local sensor data has been described in detail in the foregoing embodiments, it will not be repeated here.
  • Step S770 sending a viewing notification to the second device, so that when the second device receives the viewing notification, it stops the image degradation of the third original video image, and returns the third original video image to the first device;
  • the first device may send a viewing notification to the second device, and when receiving the viewing notification, the second device may stop reducing the resolution and/or frame rate of the third original video picture, and The third original video picture is sent to the first device.
  • a remote terminal can send a degraded video image with a resolution of 2000 pixels and a frame rate of 15FPS to the local terminal at time t+1. If it receives a notification from the local terminal that someone is watching, it can be collected at time t+2.
  • the third video picture is restored to a high-definition video picture with a resolution of 4000 pixels and a frame rate of 25FPS, and is sent to the local terminal.
  • Step S780 playing the third original video picture on the first device.
  • the first device when receiving the third original video picture sent by the second device, may play the third original video picture on the first device and display it on the screen of the first device.
  • FIG. 8 is a schematic diagram of a scene of a user returning to the scene when a video is played according to Embodiment 3 of the present disclosure.
  • the first device 302 when the first device 302 plays a degraded video image and the local user A returns, the first device 302 can detect that someone is watching and send a notification to the second device 304 , if the second device 304 After receiving a viewing notification, it can confirm that user A has returned, stop downgrading the third original video image, and send the original video image with high display quality to the first device 302, and the second device 304 can also send user A's Entry information is shown on the display.
  • the video picture quality can be restored in time when the user on the first device side returns, and the user's manual operation is not required, thereby improving the user experience.
  • FIG. 9 is a sequence diagram of a video playback method provided by the third embodiment of the present disclosure. As shown in Figure 9, the method may include the following steps:
  • Step S901 the second device collects the first original video picture
  • Step S902 the second device sends the first original video picture to the first device
  • Step S903 the first device plays the first original video picture
  • Step S904 the first device detects whether no one is watching
  • the first device collects local sensing data, determines whether there is no one at the first device by detecting whether there is human body information including human face, human voice and human body infrared data in the local sensing data, and determines whether there is no one at the first device. When there is no one, it is determined that no one is watching at the first device;
  • Step S905 the first device sends a notification that no one is watching
  • the first device generates a no-view notification, and sends the no-view notification to the second device;
  • Step S906 the second device collects the second original video picture
  • Step S907 the second device performs image degradation on the second original video image
  • the second device receives a notification that no one is watching, it performs image degradation on the second original video image to obtain a downgraded video image;
  • Step S908 the second device sends the degraded video image to the first device
  • Step S909 the first device plays the degraded video picture.
  • FIG. 10 is a sequence diagram of a server-forwarded video playback method provided by Embodiment 3 of the present disclosure. As shown in Figure 10, the method may include the following steps:
  • Step S1001 the third device collects the first original video picture
  • Step S1002 the third device sends the first original video picture to the second device
  • Step S1003 the second device forwards the first original video picture
  • the second device stores the first original video picture, and forwards the first original video picture to the first device;
  • Step S1004 the first device plays the first original video picture
  • Step S1005 the first device detects whether no one is watching
  • the first device collects local sensing data, determines whether there is no one at the first device by detecting whether there is human body information including human face, human voice and human body infrared data in the local sensing data, and determines whether there is no one at the first device. When there is no one, it is determined that no one is watching at the first device;
  • Step S1006 the first device sends a notification that no one is watching.
  • the first device generates a no-view notification, and sends the no-view notification to the second device;
  • Step S1007 the third device collects the second original video picture
  • Step S1008 the second device sends the second original video picture
  • the third device sends the second original video picture to the second device
  • Step S1009 the second device performs image degradation on the second original video image
  • the second device stores the second original video image after receiving it, and if the second device receives a notification that no one is watching, it degrades the stored second original video image to obtain a degraded video image;
  • Step S1010 the second device sends the degraded video image to the first device
  • Step S1011 the first device plays the downgraded video picture.
  • FIG. 11 is a schematic diagram of a video playback method provided in Embodiment 3 of the present disclosure.
  • position 1 may be the position of the first device
  • position 2 may be the position of the second device
  • the interested picture condition and the non-interested picture condition may be set for the video picture of the second device.
  • you can set the The interested picture condition is set to the local person
  • the non-interested picture condition is set to the local unmanned person.
  • the processing strategy for the video picture to be transmitted by the second device can be set.
  • the second device can set the non-interested picture.
  • the processing strategy is set to reduce resolution or drop frames.
  • the first device can collect local video or images, and by analyzing the video or images of the first device, it is determined that the video image of the second device conforms to the conditions of the picture of interest or the picture of non-interest. For example, if there is a local person, it can be determined that the current second The device video picture is a picture of interest. If there is no local person, it can be determined that the current second device video picture is a non-interested picture, and the first device can input the video analysis result of the person or the unmanned person into the video processing unit. After the second device collects the remote video, it can input the remote video into the video processing unit, and determine the corresponding processing strategy according to the video analysis result of the first device.
  • the resolution of the remote video can be reduced. Or discard the frames of the remote video to reduce the occupation of network transmission resources, at which time the video resolution decreases.
  • a motion detection algorithm may be used, and when no human movement is detected in the local video frame picture, the off-site video picture is determined as a non-interesting picture; a humanoid detection algorithm may also be used, when the local video frame is not detected.
  • the remote video picture When no human body shape is detected in the frame picture, the remote video picture is determined as a non-interesting picture; a face recognition algorithm can also be used, when no face image is recognized in the local video frame picture, the remote video picture is determined as Non-interesting pictures; you can also use voice synchronization input, when no human voice is detected locally, the remote video picture is determined as a non-interesting picture; you can also set the time, when the set time exceeds the preset time period Or when the playback time is long, the remote video picture is determined as a non-interesting picture.
  • FIG. 12 is a flowchart of a video playback method provided in Embodiment 4 of the present disclosure.
  • the video playback method in Embodiment 4 of the present disclosure may include the following steps:
  • Step S1210 sending the first original video picture to the first device
  • Step S1220 when receiving the unwatched notification sent by the first device, perform image degradation on the second original video image to obtain a downgraded video image, wherein the unwatched notification is that the first device is playing the first original video
  • the unwatched notification is that the first device is playing the first original video
  • the degradation of the picture includes reducing the resolution and/or reducing the frame rate
  • Step S1230 sending the degraded video images to the first device.
  • the video playback method further includes:
  • Collect the second original video picture or, receive the second original video picture collected by the third device.
  • the video playback method further includes:
  • the first original video picture is sent to the first device through the second device, and when the unwatched notification sent by the first device is received, the picture quality of the second original video picture is degraded to obtain The degraded video picture, wherein the unattended notification is a notification generated when it is determined according to local sensor data that there is no one at the first device, and the degraded video picture can be sent to the first device, and the first device can be unmanned.
  • the second device degrades the subsequently played video to reduce the video data to be transmitted, so as to better improve the occupation of network transmission resources and improve the utilization rate of network transmission resources during the video conference.
  • the above-mentioned video playback method does not need to perform too much analysis and processing on the video picture, and can save data. Process resources.
  • the above-mentioned video playback method can degrade the picture when the user temporarily leaves, reduce the data transmission bandwidth, and restore the picture quality when the user is detected to return, thereby reducing the user's manual operation on picture quality and improving the user experience.
  • the bandwidth of data transmission is reduced, and the utilization rate of network transmission resources is improved.
  • steps in the flowcharts of FIGS. 2 , 7 and 12 are sequentially displayed in accordance with the arrows, these steps are not necessarily executed in the order indicated by the arrows. Unless explicitly stated herein, the execution of these steps is not strictly limited to the order, and these steps may be performed in other orders. Moreover, at least a part of the steps in FIG. 2 , FIG. 7 and FIG. 12 may include multiple sub-steps or multiple stages, and these sub-steps or stages are not necessarily executed at the same time, but may be executed at different times. The order of execution of the sub-steps or phases is also not necessarily sequential, but may be performed alternately or alternately with other steps or at least a portion of the sub-steps or phases of the other steps.
  • Embodiment 5 of the present disclosure provides a video playback system, which may include: a first device and a second device;
  • the second device configured to send the first original video picture to the first device
  • the first device is configured to collect local sensing data when the first original video image is played, and determine whether there is no one at the first device according to the local sensing data, and when it is determined that the first device is unoccupied There is no one at the device, generate a no-view notification, and send the no-view notification to the second device;
  • the second device is further configured to downgrade the second original video image to obtain a downgraded video image when receiving the unwatched notification, and return the downgraded video to the first device A picture, wherein the picture degradation includes reducing the resolution and/or reducing the frame rate;
  • the first device is further configured to play the degraded video picture.
  • the first device is further configured to, when the degraded video image is played on the first device, if it is determined according to the local sensing data that there is a person at the first device, generate someone watching notification, and sending the someone watching notification to the second device;
  • the second device is further configured to stop performing image degradation on the third original video picture and send the third original video picture to the first device when receiving the someone watching notification;
  • the first device is further configured to play the third original video picture on the first device.
  • FIG. 13 is a schematic structural diagram of a video playback device according to Embodiment 6 of the present disclosure.
  • the video playback apparatus 1300 provided in this embodiment may include: a collection module 1301 , a determination module 1302 , a notification generation module 1303 , a notification sending module 1304 , and a playback module 1305 . in:
  • the collection module 1301 is configured to collect local sensing data when the first device plays the first original video image; the first original video image is collected by the second device at the current moment;
  • a determination module 1302, configured to determine whether there is no one at the first device according to the local sensing data
  • the notification generation module 1303 is configured to generate a notification that no one is watching if there is no one at the first device;
  • the notification sending module 1304 is configured to send the unwatched notification to the second device, so that the second device can perform image degradation on the second original video image when receiving the unwatched notification, Obtaining a downgraded video picture, and sending the downgraded video picture to the first device; the second original video picture is collected by the second device at the next moment, wherein the picture downgrading includes reducing resolution rate and/or reduced frame rate;
  • the playing module 1305 is configured to play the degraded video picture.
  • the above-mentioned collection module 1301, determination module 1302, notification generation module 1303, notification sending module 1304 and playback module 1305 can be run in the terminal as a part of the device, and the above-mentioned modules can be executed by the processor in the terminal
  • the terminal can also be a smart phone (such as Android mobile phone, iOS mobile phone, etc.), tablet computer, applause computer, Mobile Internet Devices (MID), PAD and other terminal devices.
  • the technical solutions provided by the embodiments of the present disclosure can detect whether there is no one at the first device by collecting local sensing data when the first device plays the first original video image, and determining whether there is no one at the first device according to the local sensing data. If no one is watching the first original video picture, a person watching the first original video screen will generate a no-viewing notification, and send a no-viewing notification to the second device.
  • the second device downgrades the second original video picture when receiving the notification, obtains the downgraded video picture, and sends the downgraded video picture to the first device, and the first device plays the downgraded video picture, which can be When no one is watching at the first device, notify the second device to downgrade the subsequently played video, reduce the video data that needs to be transmitted, thus reduce the occupation of network transmission resources, and avoid slow and stuck video playback. .
  • the local sensing data includes image frame data collected by the image collection device of the first device.
  • the determining module 1302 may be further configured to obtain image frame data in sequence, and perform face detection on the image frame data; when a face image is detected in the image frame data , it is determined that there is a person at the first device; when it is confirmed that no human face is detected in the acquired image frame data for a continuous preset number of times, it is determined that there is no one at the first device.
  • the determining module 1302 may be further configured to: count the silence duration in the audio data, where the silence duration is the duration during which no voice is detected in the audio data; The silence duration is compared with a preset duration threshold; when the silence duration does not exceed the duration threshold, it is determined that there is someone at the first device; when the silence duration exceeds the duration threshold, it is determined that the first device is There is no one at a device.
  • the determining module 1302 may be further configured to: determine whether the infrared data is within a preset infrared data interval; when the infrared data is within the infrared data interval, determine the infrared data There is a person at the first device; when the infrared data is not within the infrared data interval, it is determined that there is no one at the first device.
  • the second device is a terminal or a server; when the second device is the terminal, the second original video picture is the video picture collected by the second device; when the second device is the terminal When the second device is the server, the second original video image is a video image captured by a third device and sent to the second device.
  • a second notification generation module configured to generate a person viewing notification if it is determined according to the local sensor data that someone is at the first device when the degraded video image is played on the first device;
  • the second notification sending module is configured to send the someone watching notification to the second device, so that the second device stops the image degradation of the third original video picture when receiving the someone watching notification, and returning the third original video picture to the first device;
  • the second playing module is configured to play the third original video picture on the first device.
  • the above-mentioned second notification generating module, the second notification sending module and the second playing module may be run in the terminal as a part of the device, and the functions implemented by the above-mentioned modules may be executed by a processor in the terminal.
  • FIG. 14 is a schematic structural diagram of a video playback device according to Embodiment 7 of the present disclosure.
  • the video playback apparatus 1400 provided in this embodiment may include: a first sending module 1401 , a picture degradation module 1402 , and a second sending module 1403 . in:
  • the first sending module 1401 is configured to send the first original video picture to the first device
  • the picture degradation module 1402 is configured to perform picture degradation on the second original video picture when receiving the unwatched notification sent by the first device to obtain a downgraded video picture, wherein the unwatched notification is:
  • the first device collects local sensing data when playing the first original video image, and determines, according to the local sensing data, a notification generated when no one is at the first device, wherein the image Downgrading includes reducing resolution and/or reducing frame rate;
  • the second sending module 1403 is configured to send the degraded video picture to the first device.
  • first sending module 1401, picture degradation module 1402 and second sending module 1403 may run in the terminal as a part of the device, and the functions implemented by the above modules may be executed by the processor in the terminal.
  • the video playback device further includes:
  • the original video picture acquisition module is configured to collect the second original video picture; or, receive the second original video picture collected by the third device.
  • the above-mentioned original video image acquisition module may run in the terminal as a part of the device, and the functions implemented by the above-mentioned module may be executed by a processor in the terminal.
  • the video playback device further includes:
  • a picture degradation stop module configured to stop performing picture degradation on the third original video picture when receiving a notification sent by the first device for someone viewing after sending the degraded video picture to the first device , wherein the person watching notification is a notification generated when the first device determines that there is a person at the first device according to the local sensor data when playing the degraded video image;
  • a third sending module configured to send the third original video picture to the first device.
  • the above picture degradation stopping module and the third sending module may run in the terminal as a part of the apparatus, and the functions implemented by the above modules may be executed by a processor in the terminal.
  • the video playback system and video playback device provided above can be configured to execute the video playback method provided by any of the above embodiments, and have corresponding functions and beneficial effects.
  • Each module in the above video playback system and video playback device can be implemented in whole or in part by software, hardware and combinations thereof.
  • the above modules can be embedded in or independent of the processor in the computer device in the form of hardware, or stored in the memory in the computer device in the form of software, so that the processor can call and execute the operations corresponding to the above modules.
  • FIG. 15 is a schematic structural diagram of an electronic device according to Embodiment 8 of the present disclosure.
  • the electronic device includes: a processor 40 , a memory 41 , a display screen 42 with a touch function, an input device 43 , an output device 44 and a communication device 45 .
  • the number of processors 40 in the electronic device may be one or more, and one processor 40 is taken as an example in the figure.
  • the number of memories 41 in the electronic device may be one or more, and one memory 41 is taken as an example in the figure.
  • the processor 40 , the memory 41 , the display screen 42 , the input device 43 , the output device 44 and the communication device 45 of the electronic device can be connected through a bus or other means, and the connection through a bus is taken as an example in the figure.
  • the electronic device may be a computer, a mobile phone, a tablet, a projector, an interactive smart tablet, or the like. In this embodiment, description is made by taking the electronic device as an interactive smart tablet as an example.
  • the memory 41 can be configured to store software programs, computer-executable programs, and modules, such as those corresponding to the electrocardiographic signal identification method based on the multi-lead multi-structure aggregation network described in any embodiment of the present disclosure.
  • the memory 41 may mainly include a storage program area and a storage data area, wherein the storage program area may store an operating system, an application program required for at least one function; the storage data area may store data created according to the use of the device, and the like.
  • memory 41 may include high-speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other non-volatile solid state storage device.
  • memory 41 may further include memory located remotely from processor 40, which may be connected to the device through a network. Examples of such networks include, but are not limited to, the Internet, an intranet, a local area network, a mobile communication network, and combinations thereof.
  • the display screen 42 is a display screen 42 with a touch function, which can be a capacitive screen, an electromagnetic screen or an infrared screen.
  • the display screen 42 is configured to display data according to the instructions of the processor 40, and is also configured to receive a touch operation configured as the display screen 42, and send corresponding signals to the processor 40 or other devices.
  • the display screen 42 is an infrared screen, it also includes an infrared touch frame, and the infrared touch frame is arranged around the display screen 42. It can also be configured to receive infrared signals and send the infrared signals to the processor. 40 or other equipment.
  • the communication device 45 is configured to establish a communication connection with other devices, which may be a wired communication device and/or a wireless communication device.
  • the input device 43 can be configured to receive input digital or character information and generate key signal input related to user setting and function control of the electronic device, and can also be a camera configured to capture images and a pickup device to capture audio data.
  • the output device 44 may include audio equipment such as speakers. It should be noted that the specific composition of the input device 43 and the output device 44 can be set according to actual conditions.
  • the processor 40 executes various functional applications and data processing of the device by running the software programs, instructions and modules stored in the memory 41 , that is, to implement the above-mentioned multi-lead multi-structure aggregation network-based ECG signal identification method.
  • processors 40 may further implement the following operations:
  • the local sensing data includes image frame data collected by an image collection device of the first device.
  • processors 40 may further implement the following operations:
  • processors 40 further implement the following operations:
  • the silence duration is the duration in which no sound is detected in the audio data
  • the silence duration exceeds the duration threshold, it is determined that there is no one at the first device.
  • processors 40 may further implement the following operations:
  • processors 40 may further implement the following operations:
  • the second device is a terminal or a server; when the second device is the terminal, the second original video picture is the video picture collected by the second device; when the second device is the When the server is used, the second original video picture is a video picture collected by a third device and sent to the second device.
  • processors 40 may further implement the following operations:
  • processors 40 may further implement the following operations:
  • the second original video image is downgraded to obtain a downgraded video image, wherein the unwatched notification is that the first device is playing the
  • the first original video image is collected, local sensing data is collected, and it is determined according to the local sensing data that when there is no one at the first device, a notification is generated, wherein the image degradation includes reducing the resolution and/or or reduce the frame rate;
  • processors 40 may further implement the following operations:
  • the second original video picture collected by the third device is received.
  • processors 40 further implement the following operations:
  • FIG. 16 is a schematic structural diagram of an electronic device according to Embodiment 9 of the present disclosure.
  • a program product 160 according to an embodiment of the present application is described, on which computer-executable instructions are stored, the computer-executable instructions, when executed by a computer processor, are used to perform a video playback method, include:
  • the computer-executable instructions when executed by a computer processor, are used to execute program code of the following steps:
  • the local sensing data includes image frame data collected by an image collection device of the first device.
  • the computer-executable instructions when executed by a computer processor, are used to execute program code of the following steps:
  • the computer-executable instructions when executed by a computer processor, are used to execute program code of the following steps:
  • the silence duration is the duration in which no sound is detected in the audio data
  • the silence duration exceeds the duration threshold, it is determined that there is no one at the first device.
  • the computer-executable instructions when executed by a computer processor, are used to execute program code of the following steps:
  • the computer-executable instructions when executed by a computer processor, are used to execute program code of the following steps:
  • the second device is a terminal or a server; when the second device is the terminal, the second original video picture is the video picture collected by the second device; when the second device is the When the server is used, the second original video picture is a video picture collected by a third device and sent to the second device.
  • the computer-executable instructions when executed by a computer processor, are used to execute program code of the following steps:
  • the computer-executable instructions are used to execute a video playback method when executed by a computer processor, including:
  • the second original video image is downgraded to obtain a downgraded video image, wherein the unwatched notification is that the first device is playing the
  • the first original video image is collected, local sensing data is collected, and it is determined according to the local sensing data that when there is no one at the first device, a notification is generated, wherein the image degradation includes reducing the resolution and/or or reduce the frame rate;
  • the computer-executable instructions when executed by a computer processor, are used to execute program code of the following steps:
  • the second original video picture collected by the third device is received.
  • the computer-executable instructions when executed by a computer processor, are used to execute program code of the following steps:
  • a non-volatile storage medium containing computer-executable instructions provided by the embodiments of the present disclosure
  • the computer-executable instructions of the non-volatile storage medium are not limited to the operations of the video playback method as described above, and can also execute the operations of any embodiment of the present disclosure.
  • the related operations in the provided video playback method have corresponding functions and beneficial effects.
  • Nonvolatile memory may include read only memory (ROM), programmable ROM (PROM), electrically programmable ROM (EPROM), electrically erasable programmable ROM (EEPROM), or flash memory.
  • Volatile memory may include random access memory (RAM) or external cache memory.
  • RAM is available in various forms such as static RAM (SRAM), dynamic RAM (DRAM), synchronous DRAM (SDRAM), double data rate SDRAM (DDRSDRAM), enhanced SDRAM (ESDRAM), synchronous chain Road (Synchlink) DRAM (SLDRAM), memory bus (Rambus) direct RAM (RDRAM), direct memory bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM), etc.

Abstract

一种视频播放方法、系统、电子设备和存储介质。该方法包括:当第一设备播放第一原始视频画面时,采集本地传感数据(S210);根据本地传感数据,确定第一设备处是否无人(S220);当确定第一设备处无人,生成无人观看通知(S230);发送无人观看通知至第二设备,以供第二设备在接收到无人观看通知时,对第二原始视频画面进行画面降质,得到降质视频画面,并向第一设备返回降质视频画面,其中,画面降质包括降低分辨率和/或降低帧率(S240);播放降质视频画面(S250)。采用本方法能够可以在第一设备侧无人观看时,在第二设备对后续播放的视频进行画面降质,减少需要进行传输的视频数据,从而在视频会议过程中较好地改善网络传输资源的占用情况,提高网络传输资源的利用率。

Description

视频播放方法、系统、电子设备和存储介质
本公开要求于2020年11月02日提交中国专利局、申请号为202011202335.3、发明名称为“视频播放方法、系统、电子设备和存储介质”的中国专利申请的优先权,其全部内容通过引用结合在本公开中。
技术领域
本公开实施例涉及视频处理技术领域,特别是涉及一种视频播放方法、系统、电子设备和存储介质。
背景技术
目前,视频会议可以满足人们随时随地进行沟通的需要,视频会议过程中通常需要传输大量的视频、图像等信息,而且通常需要在多方用户之间进行信息传输。随着视频会议业务的增多,特别是高清视频会议业务的增多,视频会议中视频画面播放过程需要占用较多的网络传输资源,容易引起网络拥堵,造成视频画面播放出现缓慢、卡顿等现象。
常见的改善措施是视频画面发送方对待发送的视频画面增加感兴趣区域检测,然后保持感兴趣区域画质,降低非感兴趣区域的画质,从而降低视频码率,以改善网络传输资源的占用情况。但是,这种方案对视频会议过程中网络传输资源的占用情况只能起到轻微的改善作用,无法较好地改善对网络传输资源的利用。
发明内容
基于此,有必要针对上述技术问题,提供一种视频播放方法、系统、电子设备和存储介质。
第一方面,提供了一种视频播放方法,所述方法包括:
当第一设备播放第一原始视频画面时,采集本地传感数据;
根据所述本地传感数据,确定第一设备处是否无人;
当确定所述第一设备处无人,生成无人观看通知;
发送所述无人观看通知至第二设备,以供所述第二设备在接收到所述无人观看通知时,对第二原始视频画面进行画面降质,得到降质视频画面,并向所述第一设备返回所述降质视频画面,其中,所述画面降质包括降低分辨率和/或降低帧率;
播放所述降质视频画面。
第二方面,还提供了另一种视频播放方法,所述方法包括:
发送第一原始视频画面至第一设备;
当接收到所述第一设备发送的无人观看通知时,对第二原始视频画面进行画面降质,得到降质视频画面,其中,所述无人观看通知为所述第一设备在播放所述第一原始视频画面时,采集本地传感数据,根据所述本地传感数据确定所述第一设备处无人时,所生成的通知,其中,所述画面降质包括降低分辨率和/或降低帧率;
发送所述降质视频画面至所述第一设备。
第三方面,提供了一种视频播放系统,所述系统包括第一设备和第二设备;
所述第二设备,设置为发送第一原始视频画面至所述第一设备;
所述第一设备,设置为在播放所述第一原始视频画面时,采集本地传感数据,根据所述本地传感数据,确定所述第一设备处是否无人,当确定所述第一设备处无人,生成无人观看通知,并发送所述无人观看通知至所述第二设备;
所述第二设备,还设置为在接收到所述无人观看通知时,对第二原始视频画面进行画面降质,得到降质视频画面,并向所述第一设备返回所述降质视频画面,其中,所述画面降质包括降低分辨率和/或降低帧率;
所述第一设备,还设置为播放所述降质视频画面。
第四方面,提供了一种视频播放装置,所述装置包括:
采集模块,设置为当第一设备播放第一原始视频画面时,采集本地传感数据;
确定模块,设置为根据所述本地传感数据,确定第一设备处是否无人;
通知生成模块,设置为当确定所述第一设备处无人,生成无人观看通知;
通知发送模块,设置为发送所述无人观看通知至第二设备,以供所述第二设备在接收到所述无人观看通知时,对第二原始视频画面进行画面降质,得到降质视频画面,并向所述第一设备返回所述降质视频画面,其中,所述画面降质包括降低分辨率和/或降低帧率;
播放模块,设置为播放所述降质视频画面。
第五方面,还提供了另一种视频播放装置,所述装置包括:
第一发送模块,设置为发送第一原始视频画面至第一设备;
画面降质模块,设置为当接收到所述第一设备发送的无人观看通知时,对第二原始视频画面进行画面降质,得到降质视频画面,其中,所述无人观看通知为所述第一设备在播放所述第一原始视频画面时,采集本地传感数据,根据所述本地传感数据确定所述第一设备处无人时,所生成的通知,其中,所述画面降质包括降低分辨率和/或降低帧率;
第二发送模块,设置为发送所述降质视频画面至所述第一设备。
第六方面,提供了一种电子设备,包括:存储器、一个或多个处理器;
所述存储器,设置为存储一个或多个程序;
当所述一个或多个程序被所述一个或多个处理器执行,使得所述一个或多个处理器实现如下操作:
当第一设备播放第一原始视频画面时,采集本地传感数据;
根据所述本地传感数据,确定第一设备处是否无人;
当确定所述第一设备处无人,生成无人观看通知;
发送所述无人观看通知至第二设备,以供所述第二设备在接收到所述无人观看通知时,对第二原始视频画面进行画面降质,得到降质视频画面,并向所述第一设备返回所述降质视频画面,其中,所述画面降质包括降低分辨率和/或降低帧率;
播放所述降质视频画面。
第七方面,提供了一种包含计算机可执行指令的非易失性存储介质,其上存储有计算机程序,所述计算机程序被处理器执行时实现以下步骤:
当第一设备播放第一原始视频画面时,采集本地传感数据;
根据所述本地传感数据,确定第一设备处是否无人;
当确定所述第一设备处无人,生成无人观看通知;
发送所述无人观看通知至第二设备,以供所述第二设备在接收到所述无人观看通知时,对第二原始视频画面进行画面降质,得到降质视频画面,并向所述第一设备返回所述降质视频画面,其中,所述画面降质包括降低分辨率和/或降低帧率;
播放所述降质视频画面。
上述视频播放方法、系统、电子设备和存储介质,通过当第一设备播放第一原始视频画面时,采集本地传感数据,根据本地传感数据确定第一设备处是否无人,可以检测第一设备侧是否无人观看,当确定第一设备处无人时,生成无人观看通知,发送无人观看通知至第二设备,可以在第一设备侧无人观看时通知第二设备,第二设备在接收到通知时对第二原始视频画面进行画面降质,得到降质视频画面,并向第一设备返回降质视频画面,第一设备播放降质视频画面,可以在第一设备侧无人观看时,在第二设备对后续播放的视频进行画面降质,减少需要进行传输的视频数据,从而在视频会议过程中较好地改善网络传输资源的占用情况,提高网络传输资源的利用率。
进一步地,相比于相关技术中在第二设备视频画面中检测感兴趣区域和降低非感兴趣区域画质的方法,上述视频播放方法无需对视频画面进行太多的分析和处理,可以节约数据处理资源。
进一步地,上述视频播放方法可以在用户临时离开时进行画面降质,减少数据传输带宽,在检测到用户返回时恢复画质,从而减少用户对于画质的手动操作,提升用户体验,而且,可以在第一设备侧用户未察觉的情况下,减少数据传输带宽,提高网络传输资源的利用率。
附图说明
图1是本公开实施例一提供的一种视频播放方法的应用环境的示意图;
图2是本公开实施例一提供的一种视频播放方法的流程图;
图3是本公开实施例一提供的一种视频播放方法的应用环境示意图;
图4是本公开实施例一提供的一种视频播放时用户离场的场景示意图;
图5是本公开实施例一提供的一种多用户视频播放的场景示意图;
图6是本公开实施例二提供的一种视频播放方法的应用环境示意图;
图7是本公开实施例三提供的一种视频播放方法的流程图;
图8是本公开实施例三提供的一种视频播放时用户返场的场景示意图;
图9是本公开实施例三提供的一种视频播放方法的时序图;
图10是本公开实施例三提供的一种服务器转发视频播放方法的时序图;
图11是本公开实施例三提供的一种视频播放方法的示意图;
图12是本公开实施例四提供的一种视频播放方法的流程图;
图13是本公开实施例六提供的一种视频播放装置的结构示意图;
图14是本公开实施例七提供的一种视频播放装置的结构示意图;
图15是本公开实施例八提供的一种电子设备的结构示意图;
图16是本公开实施例九提供的一种电子设备的结构示意图。
具体实施方式
为了使本公开实施例的目的、技术方案及优点更加清楚明白,以下结合附图及实施例,对本公开实施例进行进一步详细说明。应当理解,此处描述的具体实施例仅仅用以解释本公开实施例,并不用于限定本公开实施例。
图1是本公开实施例一提供的一种视频播放方法的应用环境的示意图。参考图1,在进行视频会议过程中,可以有至少一个本地设备102和至少一个异地设备104,其中,本地设备102和异地设备104可以包括但不限于各种配置有摄像头的电脑、手机、平板或投影仪,本地设备102和异地设备104通过有线或无线网络相通信,可以直接相通信,也可以通过服务器相通信,本地设备102的显示屏可以显示异地用户B和C的视频画面,异地设备104的显示屏上可以至少显示本地用户A的视频画面,本地设备102还可以显示异地设备104所共享的视频、图片和文档,异地设备104也可以显示本地设备102所共享的视频、图片和文档。
以下针对视频会议场景,以本地设备102对用户B的视频画面进行播放为 例,对视频播放过程进行说明。在本地设备102与用户B所使用的异地设备104建立通信连接后,异地设备104可以通过摄像头对用户B进行实时拍摄,获取到用户B的视频,并将用户B的视频按照设定的分辨率和帧率进行调整,生成具有固定分辨率和固定帧率的用户B的视频画面。为确保视频画面能够高效、可靠地传输至本地设备102,异地设备104可以使用视频编码器对用户B的视频画面进行编码,视频编码器按照预设码率对视频画面进行数据压缩,生成数据量相对较小的视频编码数据,异地设备104可以将视频编码数据传输至本地设备102。本地设备102在接收到视频编码数据后,可以通过视频解码器对其进行解码,还原用户B的视频画面,供本地设备102播放,本地设备102上可以实时显示用户B的视频画面。
需要说明的是,当异地设备104共享视频、图片或文档时,可以省略对用户B的视频拍摄过程,视频编码器可以直接对异地设备104上显示的画面进行编码,并将编码后的数据传输至本地设备102,本地设备102在对其进行解码后,可以实时显示异地设备104所共享的视频、图片或文档。
异地设备104对用户A的视频画面进行播放,以及共享本地设备视频、图片或文档的过程,可以为上述过程的逆过程,在此不再赘述。
对于高清视频会议场景,为了播放高清视频画面,异地设备104可以生成分辨率和/或帧率较高的视频画面,传输至本地设备102进行播放。在视频编码码率不变的情况下,视频播放过程中需要传输的数据量较大,占用较多的网络传输资源。相关技术中通常通过在待发送的视频画面中检测感兴趣区域,保持感兴趣区域画质,降低非感兴趣区域画质来改善对网络传输资源的占用,但是这种方法仍然无法较好地改善对于网络传输资源的占用情况,难以提高网络传输资源的利用率。
实施例一
图2是本公开实施例一提供的一种视频播放方法的流程图。
图3是本公开实施例一提供的一种视频播放方法的应用环境示意图。如图3所示,用户通过至少一个第一设备302和至少一个第二设备304进行视频会议,其中,第一设备302可以为本地终端设备,第二设备304可以为异地终端设备。 本实施例提供的视频播放方法可以由终端设备执行,该终端设备可以通过软件和/或硬件的方式实现,其中,软件的实现方式可以为在硬件上运行软件,该终端设备可以是由两个或多个物理实体构成,也可以是由一个物理实体构成。该终端设备可以是电脑、手机、平板或投影仪等。
可选地,参考图2,本公开实施例一的视频播放方法,以应用于图3中的第一设备302为例,可以包括以下步骤:
步骤S210,当第一设备播放第一原始视频画面时,采集本地传感数据。
在该实施例中,原始视频画面可以为高显示质量的视频画面,第一原始视频画面可以为第二设备在当前时刻采集到的原始视频画面。
其中,本地传感数据可以为第一设备采集到的数据,其中,第一设备可以通过图像采集设备、音频采集设备和红外传感设备中的至少一种采集本地传感数据。
在具体实现中,在进行视频会议过程中,第二设备可以采集异地用户的视频,得到高显示质量的原始视频画面,通过预设一定的采集周期,可以将当前时刻采集到的原始视频画面确定为第一原始视频画面,经过一定的采集周期到达当前时刻的下一时刻,可以将下一时刻采集到的原始视频画面确定为第二原始视频画面,依此类推。当前时刻,第二设备可以将高显示质量的第一原始视频画面发送至第一设备,该第一设备在接收到第一原始视频画面后,播放第一原始视频画面,将第一原始视频画面显示在第一设备的屏幕上。第一设备可以配置有图像采集设备、音频采集设备和红外传感设备中的至少一种,当第一设备播放第一原始视频画面时,第一设备可以通过图像采集设备采集本地的图像帧数据,将图像帧数据作为本地传感数据,还可以通过音频采集设备来采集第一设备周围的音频数据,将采集到的音频数据作为本地传感数据,还可以通过红外传感器来采集第一设备周围的红外数据,将采集到的红外数据作为本地传感数据。
例如,第一设备可以配置有前置摄像头,在播放第一原始视频画面时,可以通过前置摄像头拍摄第一设备显示屏前方一定范围内的视频,得到图像帧数据。
需要说明的是,上述原始视频画面(包括第一原始视频画面和第二原始视频画面)可以不限于是采集到的异地用户的视频画面,原始视频画面还可以为第二设备所共享的视频、图片和文档的画面。
步骤S220,根据本地传感数据,确定第一设备处是否无人。
步骤S230,当确定第一设备处无人时,则生成无人观看通知。
在具体实现中,第一设备可以根据本地传感数据确定第一设备处有人或无人的情况,并在无人时生成无人观看通知。
在一种可选的实施方式中,第一设备可以通过图像采集设备采集本地视频,从本地视频中截取一帧图像或连续多帧图像,作为图像帧数据,第一设备还可以通过图像采集设备直接采集本地的图像帧数据。在采集到图像帧数据后,第一设备可以检测图像帧数据中是否有人脸,若检测到人脸,则可以确定第一设备处有人,若未检测到人脸,则可以确定第一设备处无人,并在第一设备生成无人观看通知。例如,在第一设备播放第一原始视频画面时,可以在本地视频中截取连续的N帧图像,若在任意一帧中检测到人脸,则可以确定第一设备处有人,若在N帧图像中均未检测到人脸,则可以确定第一设备处无人,并生成无人观看通知。
在另一种可选的实施方式中,若本地传感数据为音频数据,则第一设备可以统计音频数据中持续无人说话的时间长度(静默时长),将时间长度与预设的时长阈值相比较,若时间长度未超过时长阈值,则可以确定第一设备处有人,若时间长度超过时长阈值,则可以确定第一设备处无人,并在第一设备生成无人观看通知。其中,第一设备可以将采集到的音频数据与预设的音频范围相比较,若采集到的音频数据不在预设的音频范围内,则可以判定无人说话。例如,在第一设备播放第一原始视频画面时,可以开始以1秒为间隔采集第一设备周围的音频,并识别音频中是否有人声,若无人声则开始计时,并统计无人声的时间长度,若时间长度未超过时长阈值10秒,则可以确定第一设备处有人,否则,若时间长度超过10秒,则可以确定第一设备处无人,并在第一设备生成无人观看通知。还可以在第一设备设置音频范围500-2000Hz(赫兹),将采集到的音频数据与其相比较,若音频数据为400Hz,不在上述音频范围内,则可以判定 无人说话。
在另一种可选的实施方式中,若本地传感数据为红外数据,第一设备可以将红外数据与预设的红外数据区间相比较,若红外数据在该红外数据区间以内,则可以确定第一设备处有人,若红外数据在红外数据区间以外,则可以确定第一设备处无人,并在第一设备生成无人观看通知。例如,在第一设备播放第一原始视频画面时,可以通过红外测温仪测量第一设备周围的温度,若测量到的温度在人体温度区间内,则可以确定第一设备处有人,否则,若测量到的温度不在人体温度区间内,则可以确定第一设备处无人,并生成无人观看通知。
步骤S240,发送无人观看通知至第二设备,以供第二设备在接收到无人观看通知时,对第二原始视频画面进行画面降质,得到降质视频画面,并向第一设备返回降质视频画面,其中,画面降质包括降低分辨率和/或降低帧率。
步骤S250,播放降质视频画面。
其中,画面降质可以为降低画面显示质量的操作,包括降低分辨率和/或降低帧率。降质视频画面可以为画面显示质量降低的视频画面。
其中,第二原始视频画面可以为第二设备在当前时刻的下一时刻采集到的原始视频画面。
其中,第二设备可以为终端设备,还可以为服务器设备。
在具体实现中,第一设备可以将无人观看通知发送至第二设备,该第二设备在接收到无人观看通知时,可以在当前时刻的下一时刻对采集到的第二原始视频画面进行画面降质,得到降质视频画面,第二设备可以将降质视频画面发送至第一设备,该第一设备在接收到第二设备发送的降质视频画面时,可以播放降质视频画面,将降质视频画面显示在第一设备的屏幕上。
例如,第二设备可以在接收到无人观看通知时,将下一时刻采集到的高质量视频画面降低分辨率,生成分辨率较低的降质视频画面,发送给第一设备进行播放;第二设备还可以将下一时刻采集到的高质量视频画面降低帧率,生成帧率较低的降质视频画面,发送给第一设备进行播放。
可选地,若第二设备的画面降质是将第二原始视频画面的帧率降低至0,则可以不向第一设备传输降质视频画面,此种情况下第一设备显示静止的视频画 面。
需要说明的是,若第一设备检测到第一原始视频画面有人观看,则无需发送通知,第二设备可以在下一时刻直接将采集到的第二原始视频画面发送至第一设备,第一设备可以对第二原始视频画面进行播放。若第一设备检测到第一原始视频画面由无人观看转为有人观看,还可以向第二设备发送有人观看通知,第二设备在接收到有人观看通知时,可以在下一时刻停止画面降质,直接将采集到的第二原始视频画面发送至第一设备,以供第一设备播放。
图4是本公开实施例一提供的一种视频播放时用户离场的场景示意图。如图4所示,可以设定第一设备302当前时刻播放的原始视频画面分辨率为4000像素,帧率为25帧每秒(Frame Per Second,简称为FPS),在第一设备302播放过程中可以实时检测第一设备302是否存在人体信息,例如,第一设备302可以采集本地视频,通过检测本地视频中是否有人体移动、人体形状、人脸图像来判断第一设备302是否存在人体信息,还可以通过在第一设备302周围一定范围内是否能检测到人体红外数据或人声来判断第一设备302是否存在人体信息。若第一设备302未检测到人体信息,则可以判定当前时刻第一设备302无人,原始视频画面无人观看,若检测到人体信息,则可以判定第一设备302有人,原始视频画面有人观看;第一设备302还可以配置有触摸屏,通过设置一个时间阈值,当触摸屏未接收到用户的触屏操作、且持续时间超过该时间阈值时,可以判定当前时刻的原始视频画面无人观看,否则可以判定为有人观看。如图4所示,当本地用户A离开、第一设备302检测到当前时刻的原始视频画面无人观看时,可以通知第二设备304,第二设备304在接收到通知时可以在显示屏上显示本地用户A已离开,并对下一时刻的原始视频画面进行画面降质,例如,可以将分辨率降低至2000像素、或将帧率降低至15FPS,并将降质视频画面发送至第一设备302进行显示,例如,第一设备302可以显示低分辨率的异地用户B图像,第二设备304还可以将帧率降至0,不向第一设备302传输视频画面,此种情况下第一设备302显示静止的视频画面。
图5是本公开实施例一提供的一种多用户视频播放的场景示意图。如图5所示,该场景下包括第一设备302和至少两个第二设备304,第一设备302有本 地用户A,第二设备304有异地用户B和C,当本地用户A离开时,第一设备302可以发送无人观看通知给所有第二设备304,第二设备304在接收到无人观看通知时,可以在显示屏上显示本地用户A已离开,并正常显示未离开的用户B或C。第二设备304还可以根据不同的分辨率/帧率对各自的原始视频画面进行画面降质,生成降质视频画面,并将降质视频画面分别发送给第一设备302,在第一设备302上显示异地用户B和C的降质视频画面。
本公开实施例的技术方案,通过当第一设备播放第一原始视频画面时,采集本地传感数据,根据本地传感数据确定第一设备处是否无人,可以检测第一设备侧是否无人观看,当确定第一设备处无人时,生成无人观看通知,发送无人观看通知至第二设备,可以在第一设备侧无人观看时通知第二设备,第二设备在接收到通知时对第二原始视频画面进行画面降质,得到降质视频画面,并向第一设备返回降质视频画面,第一设备播放降质视频画面,可以在第一设备侧无人观看时,在第二设备对后续播放的视频进行画面降质,减少需要进行传输的视频数据,从而在视频会议过程中较好地改善网络传输资源的占用情况,提高网络传输资源的利用率。
进一步地,相比于相关技术中在第二设备视频画面中检测感兴趣区域和降低非感兴趣区域画质的方法,上述视频播放方法无需对视频画面进行太多的分析和处理,可以节约数据处理资源。
进一步地,上述视频播放方法可以在用户临时离开时进行画面降质,减少数据传输带宽,在检测到用户返回时恢复画质,从而减少用户对于画质的手动操作,提升用户体验,而且,可以在第一设备侧用户未察觉的情况下,减少数据传输带宽,提高网络传输资源的利用率。
可选地,所述本地传感数据包括第一设备的图像采集设备采集的图像帧数据。
其中,图像采集设备可以为采集第一设备周围视频或图像的设备,可以包括但不限于是各种摄像头和摄像机。
在具体实现中,第一设备可以配置有图像采集设备,当第一设备播放第一 原始视频画面时,第一设备可以通过图像采集设备采集本地视频,从本地视频中截取一帧图像或连续多帧图像,作为图像帧数据,第一设备还可以通过图像采集设备直接采集本地的图像帧数据。例如,第一设备可以配置有前置摄像头,在播放第一原始视频画面时,通过前置摄像头拍摄第一设备显示屏前方一定范围内的视频,通过截取视频中的图像帧得到图像帧数据,还可以通过前置摄像头拍摄第一设备显示屏前方一定范围内的图像,直接得到图像帧数据。
根据本公开实施例的技术方案,本地传感数据包括第一设备的图像采集设备采集的图像帧数据,可以便于检测本地无人情况。
可选地,所述步骤S220,包括:
依次获取图像帧数据,并对图像帧数据进行人脸检测;当在图像帧数据中检测到人脸图像时,确定第一设备处有人;当连续预设次数在获取的所述图像帧数据中未检测到人脸,确定第一设备处无人。
其中,人脸检测可以为检测图像中是否有人脸图像。
在具体实现中,第一设备可以从本地视频中依次获取图像帧数据,并检测图像帧数据中是否有人脸,若检测到人脸图像,则可以确定第一设备处有人,若未检测到人脸图像,则可以确定第一设备处无人。
例如,在播放第一原始视频画面时,可以在本地视频中截取连续的N帧图像,若在任意一帧中检测到人脸,则可以确定第一设备处有人,若在N帧图像中均未检测到人脸图像,则可以确定第一设备处无人。为提高检测的准确性,还可以在N帧图像中的至少M(M≤N)帧中检测到人脸图像时,确定第一设备处有人,若在小于M帧中检测到人脸图像,则可以确定第一设备处无人。
需要说明的是,人脸检测还可以为检测图像中的人脸是否为指定的人脸图像,例如,可以检测图像帧数据中的人脸图像是否与本地用户A的人脸图像相匹配,若匹配,则可以确定第一设备处有人,若不匹配,则确定第一设备处无人。
根据本公开实施例的技术方案,通过依次获取图像帧数据,并对图像帧数据进行人脸检测,可以检测到图像帧数据中是否有人脸图像,当在图像帧数据 中检测到人脸图像时,确定第一设备处有人,当连续预设次数在获取的所述图像帧数据中未检测到人脸图像时,确定第一设备处无人,可以根据人脸检测的结果确定第一设备处是否无人,以便在无人时降低显示画面质量,减少需要进行传输的视频数据,从而在视频会议过程中较好地改善网络传输资源的占用情况,提高网络传输资源的利用率。
可选地,所述本地传感数据还包括第一设备的音频采集设备采集的音频数据,所述步骤S220,还包括:
统计音频数据中的静默时长,其中,静默时长为音频数据中未检测到有人发声的时长;将静默时长与预设的时长阈值相比较;当静默时长未超过时长阈值时,确定第一设备处有人;当静默时长超过时长阈值时,确定第一设备处无人。
其中,音频采集设备可以为采集第一设备周围音频的设备,可以包括但不限于是各种传声器。
其中,音频数据可以为在第一设备周围采集到的声音。
其中,静默时长可以为音频数据中无人说话的时间长度。
在具体实现中,第一设备可以配置有音频采集设备,当第一设备播放第一原始视频画面时,第一设备通过音频采集设备采集本地的音频数据,将音频数据作为本地传感数据。第一设备还可以将音频数据与人声的音频范围相比较,若本地音频不在人声的音频范围内,则可以确定无人说话,第一设备可以统计音频数据中持续无人说话的时间长度,得到静默时长,将静默时长与预设的时长阈值相比较,若静默时长未超过时长阈值,则可以确定第一设备处有人,若静默时长超过时长阈值,则可以确定第一设备处无人。
例如,第一设备可以配置有麦克风,在第一设备播放第一原始视频画面时,通过麦克风采集第一设备周围一定范围内的音频数据,还可以设置音频范围500-2000Hz,将采集到的音频数据与音频范围相比较,若音频数据为400Hz,不在上述音频范围内,则可以判定为静默。在第一设备播放第一原始视频画面时,可以开始以1秒为间隔采集第一设备周围的音频数据,并识别音频数据是否为 静默,若为静默则开始计时,并统计静默时长,若静默时长未超过时长阈值10秒,则可以确定第一设备处有人,否则,若静默时长超过10秒,则可以确定第一设备处无人。
需要说明的是,还可以检测音频数据是否为指定用户的音频数据,例如,可以检测音频数据是否与本地用户A的音频数据相匹配,若匹配,则可以确定本地用户A未静默,若不匹配,则可以确定本地用户A静默,并通过统计本地用户A的静默时长来确定第一设备处是否无人。
根据本公开实施例的技术方案,通过统计音频数据中的静默时长,将静默时长与预设的时长阈值相比较,当静默时长未超过时长阈值时,确定第一设备处有人,当静默时长超过时长阈值时,确定第一设备处无人,可以根据静默时长的统计结果确定第一设备处是否无人,以便在无人时降低显示画面质量,减少需要进行传输的视频数据,从而在视频会议过程中较好地改善网络传输资源的占用情况,提高网络传输资源的利用率。
可选地,所述本地传感数据还包括第一设备的红外传感设备采集的红外数据,所述步骤S220,还包括:
判断红外数据是否处于预设的红外数据区间内;当红外数据处于红外数据区间内时,确定第一设备处有人;当红外数据未处于红外数据区间内时,确定第一设备处无人。
其中,红外传感设备可以为采集第一设备周围红外数据的设备,可以包括但不限于是各种红外测温仪和红外成像仪。
其中,红外数据可以为在第一设备周围采集到的红外数据。
在具体实现中,第一设备可以配置有红外传感设备,当第一设备播放第一原始视频画面时,第一设备通过红外传感设备采集本地的红外数据。第一设备还可以将采集到的红外数据与预设的红外数据区间相比较,若红外数据在红外数据区间以内,则可以确定第一设备处有人,若红外数据在红外数据区间以外,则可以确定第一设备处无人。
例如,第一设备可以配置有红外测温仪,在第一设备播放第一原始视频画 面时,通过红外测温仪测量第一设备周围一定范围内的温度,将其作为红外数据,若测量到的温度在人体温度区间(红外数据区间)内,则可以确定第一设备处有人,否则,若测量到的温度不在人体温度区间内,则可以确定第一设备处无人。
根据本公开实施例的技术方案,通过判断红外数据是否处于预设的红外数据区间内,当红外数据处于红外数据区间内时,确定第一设备处有人,当红外数据未处于红外数据区间内时,确定第一设备处无人,可以根据红外数据所处的区间确定第一设备处是否无人,以便在无人时降低显示画面质量,减少需要进行传输的视频数据,从而在视频会议过程中较好地改善网络传输资源的占用情况,提高网络传输资源的利用率。
实施例二
图6是本公开实施例二提供的一种视频播放方法的应用环境示意图。如图所示,用户通过第一设备302、第二设备304和第三设备304进行视频会议,其中,第一设备302可以为本地终端设备,第二设备304可以为服务器,第三设备306可以为异地终端设备。其中,服务器可以是为视频播放提供临时存储和转发的服务器,可以用独立的服务器或者是多个服务器组成的服务器集群来实现,还可以为云服务器。
可选地,参考图6,本公开实施例二的视频播放方法,可以包括:
第二设备为终端或服务器;当第二设备为终端时,第二原始视频画面为第二设备采集到的视频画面;当第二设备为服务器时,第二原始视频画面为由第三设备采集、并发送至第二设备的视频画面。
在具体实现中,当第二设备为终端时,如图3-5所示,第二设备304可以为异地终端,通过采集异地用户视频,得到高显示质量的原始视频画面,其中包括第二原始视频画面。当第二设备为服务器时,如图6所示,第三设备306可以为异地终端,在采集异地用户视频,得到高显示质量的原始视频画面后,第三设备306可以将原始视频画面发送至第二设备304,第二设备304对原始视频画面进行存储,并转发给第一设备302,其中,原始视频画面可以包括第二原始 视频画面。
当第二设备为服务器时,异地终端可以将采集到的原始视频画面发送至服务器,服务器对原始视频画面进行存储,并转发给本地终端。异地终端可以在当前时刻将第一原始视频画面发送给服务器,在下一时刻将第二原始视频画面发送给服务器。在当前时刻,服务器可以将高显示质量的第一原始视频画面发送至本地终端,本地终端在接收到第一原始视频画面后,对第一原始视频画面进行播放,并通过采集本地传感数据来检测第一原始视频画面是否有人观看,若检测到无人观看,则可以向服务器发送无人观看通知,服务器在接收到无人观看通知时,可以在当前时刻的下一时刻将存储的第二原始视频画面降低分辨率和/或降质帧率,得到降质视频画面,服务器可以将降质视频画面发送给本地终端进行播放。
其中,本地终端可以有本地用户A,异地终端可以有异地用户B和C,本地终端可以从服务器获取异地终端的视频画面,当本地用户A离开时,本地终端可以通知服务器,服务器在接收到通知时可以发送信息给异地终端,异地终端可以在显示屏上显示本地用户A已离开,并正常显示未离开的用户B或C,服务器还可以对服务器中存储的异地用户B和C的视频画面进行画面降质,并将降质视频画面发送给本地终端,本地终端上可以显示异地用户B和C的降质视频画面。
根据本公开实施例的技术方案,第二设备为终端或服务器,当第二设备为终端时,第二原始视频画面为第二设备采集到的视频画面,当第二设备为服务器时,第二原始视频画面为由第三设备采集、并发送至第二设备的视频画面,可以由服务器执行对第二原始视频画面的画面降质,便于服务器进行集中式管理,提高视频处理速度,降低视频播放延时。
实施例三
图7是本公开实施例三提供的一种视频播放方法的流程图。可选地,以应用于图3-6中的第一设备302为例,本公开实施例三的视频播放方法,可以包括以下步骤:
步骤S710,当第一设备播放第一原始视频画面时,采集本地传感数据;
步骤S720,根据本地传感数据,确定第一设备处是否无人;
步骤S730,当确定第一设备处无人时,生成无人观看通知;
步骤S740,发送无人观看通知至第二设备,以供第二设备在接收到无人观看通知时,对第二原始视频画面进行画面降质,得到降质视频画面,并向第一设备返回降质视频画面,其中,画面降质包括降低分辨率和/或降低帧率;
步骤S750,播放降质视频画面;
步骤S760,当在第一设备上播放降质视频画面时,若根据本地传感数据,确定第一设备处有人,则生成有人观看通知;
在具体实现中,当第一设备在播放降质视频画面时,可以通过图像采集设备、音频采集设备和红外传感设备中的至少一种来采集本地传感数据,通过检测本地传感数据中是否存在包括人脸、人声和人体红外数据在内的人体信息,可以确定第一设备周围一定范围内是否有人,第一设备可以在有人时生成有人观看通知。由于根据本地传感数据确定第一设备处有人的过程在前述实施例中已有详细说明,在此不再赘述。
步骤S770,发送有人观看通知至第二设备,以供第二设备在接收到有人观看通知时,停止对第三原始视频画面进行画面降质,并向第一设备返回第三原始视频画面;
在具体实现中,第一设备可以将有人观看通知发送至第二设备,第二设备在接收到有人观看通知时,可以停止对第三原始视频画面进行降低分辨率和/或降低帧率,并将第三原始视频画面发送给第一设备。
例如,异地终端可以在t+1时刻向本地终端发送分辨率2000像素、帧率15FPS的降质视频画面,若接收到本地终端反馈的有人观看通知,则可以在t+2时刻,将采集到的第三视频画面恢复至分辨率4000像素、帧率25FPS的高清视频画面,发送给本地终端。
步骤S780,在第一设备上播放第三原始视频画面。
在具体实现中,第一设备在接收到第二设备发送的第三原始视频画面时,可以在第一设备上播放第三原始视频画面,显示在第一设备的屏幕上。
图8是本公开实施例三提供的一种视频播放时用户返场的场景示意图。如图8所示,当第一设备302播放降质视频画面、且本地用户A返回时,第一设备302可以检测到有人观看,并发送有人观看通知给第二设备304,第二设备304若接收到有人观看通知,则可以确认用户A已经返回,并停止对第三原始视频画面进行降质,向第一设备302发送高显示质量的原始视频画面,第二设备304还可以将用户A的返场信息显示在显示屏上。
根据本公开实施例的技术方案,当在第一设备上播放降质视频画面时,若根据本地传感数据,确定第一设备处有人,则生成有人观看通知,发送有人观看通知至第二设备,并播放第三原始视频画面,可以在第一设备侧用户返回时,及时恢复视频画面质量,且无需用户手动操作,提升用户体验。
为了更清晰阐明本公开实施例提供的技术方案,以下将结合图9和10对该方案进行说明,图9是本公开实施例三提供的一种视频播放方法的时序图。如图9所示,该方法可以包括以下步骤:
步骤S901,第二设备采集第一原始视频画面;
步骤S902,第二设备将第一原始视频画面发送至第一设备;
步骤S903,第一设备播放第一原始视频画面;
步骤S904,第一设备检测是否无人观看;
第一设备采集本地传感数据,通过检测本地传感数据中是否存在包括人脸、人声和人体红外数据在内的人体信息,来判断第一设备处是否无人,并在第一设备处无人时,确定第一设备处无人观看;
步骤S905,第一设备发送无人观看通知;
第一设备生成无人观看通知,将无人观看通知发送给第二设备;
步骤S906,第二设备采集第二原始视频画面;
步骤S907,第二设备对第二原始视频画面进行画面降质;
若第二设备接收到无人观看通知,则对第二原始视频画面进行画面降质,得到降质视频画面;
步骤S908,第二设备发送降质视频画面至第一设备;
步骤S909,第一设备播放降质视频画面。
图10是本公开实施例三提供的一种服务器转发视频播放方法的时序图。如图10所示,该方法可以包括以下步骤:
步骤S1001,第三设备采集第一原始视频画面;
步骤S1002,第三设备将第一原始视频画面发送至第二设备;
步骤S1003,第二设备转发第一原始视频画面;
第二设备存储第一原始视频画面,并将第一原始视频画面转发至第一设备;
步骤S1004,第一设备播放第一原始视频画面;
步骤S1005,第一设备检测是否无人观看;
第一设备采集本地传感数据,通过检测本地传感数据中是否存在包括人脸、人声和人体红外数据在内的人体信息,来判断第一设备处是否无人,并在第一设备处无人时,确定第一设备处无人观看;
步骤S1006,第一设备发送无人观看通知。
第一设备生成无人观看通知,将无人观看通知发送给第二设备;
步骤S1007,第三设备采集第二原始视频画面;
步骤S1008,第二设备发送第二原始视频画面;
第三设备将第二原始视频画面发送至第二设备;
步骤S1009,第二设备对第二原始视频画面进行画面降质;
第二设备在接收到第二原始视频画面后对其进行存储,若第二设备接收到无人观看通知,则对存储的第二原始视频画面进行画面降质,得到降质视频画面;
步骤S1010,第二设备发送降质视频画面至第一设备;
步骤S1011,第一设备播放降质视频画面。
为了便于本领域技术人员深入理解本公开实施例,以下将结合一个具体示例进行说明。
图11是本公开实施例三提供的一种视频播放方法的示意图。如图11所示,位置1可以为第一设备所在位置,位置2可以为第二设备所在位置,为第二设 备的视频画面设定感兴趣画面条件和非感兴趣画面条件,例如,可以将感兴趣画面条件设置为本地有人,将非感兴趣画面条件设置为本地无人,相应地,可以设置对于第二设备待传输视频画面的处理策略,例如,可以将第二设备非感兴趣画面的处理策略设置为降低分辨率或丢弃帧画面。第一设备可以采集本地的视频或图像,通过分析第一设备视频或图像,确定第二设备视频画面符合感兴趣画面条件或非感兴趣画面条件,例如,若本地有人,则可以判定当前第二设备视频画面为感兴趣画面,若本地无人,则可以判定当前第二设备视频画面为非感兴趣画面,第一设备可以将有人或无人的视频分析结果输入视频处理单元。第二设备在采集到异地视频后,可以将异地视频输入视频处理单元,并根据第一设备的视频分析结果确定相应的处理策略,例如,若本地有人,当前异地视频画面为感兴趣画面,则可以不对异地视频进行画面降质,保持视频清晰度不变,确保第一设备可以显示优质的视频画面,若本地无人,当前异地视频画面为非感兴趣画面,则可以降低异地视频的分辨率或丢弃异地视频的帧画面,减少对网络传输资源的占用,此时视频清晰度下降。
可选地,该实施例可以使用移动侦测算法,当在本地视频帧画面中未检测到人体移动时,将异地视频画面确定为非感兴趣画面;还可以使用人形检测算法,当在本地视频帧画面中未检测到人体形状时,将异地视频画面确定为非感兴趣画面;还可以使用人脸识别算法,当在本地视频帧画面中未识别到人脸图像时,将异地视频画面确定为非感兴趣画面;还可以使用声音同步输入,当在本地未检测到人声时,将异地视频画面确定为非感兴趣画面;还可以进行时间设定,当设定的时间超过预设的时段或播放时长时,将异地视频画面确定为非感兴趣画面。
实施例四
图12是本公开实施例四提供的一种视频播放方法的流程图。可选地,以应用于图3-6中的第二设备304为例,本公开实施例四的视频播放方法,可以包括以下步骤:
步骤S1210,发送第一原始视频画面至第一设备;
步骤S1220,当接收到第一设备发送的无人观看通知时,对第二原始视频画面进行画面降质,得到降质视频画面,其中,无人观看通知为第一设备在播放第一原始视频画面时,采集本地传感数据,根据本地传感数据确定第一设备处无人时,所生成的通知,其中,画面降质包括降低分辨率和/或降低帧率;
步骤S1230,发送降质视频画面至第一设备。
可选地,所述视频播放方法,还包括:
采集第二原始视频画面;或,接收第三设备所采集的第二原始视频画面。
可选地,所述视频播放方法,还包括:
在发送降质视频画面至第一设备之后,当接收到第一设备发送的有人观看通知时,停止对第三原始视频画面进行画面降质,其中,有人观看通知为第一设备在播放降质视频画面时,根据本地传感数据确定第一设备处有人时,所生成的通知;发送第三原始视频画面至第一设备。
由于第二设备304的处理过程在前述实施例中已有详细说明,在此不再赘述。
根据本公开实施例的技术方案,通过第二设备发送第一原始视频画面至第一设备,当接收到第一设备发送的无人观看通知时,对第二原始视频画面进行画面降质,得到降质视频画面,其中,无人观看通知为根据本地传感数据确定第一设备处无人时,所生成的通知,可以发送降质视频画面至第一设备,可以在第一设备侧无人观看时,在第二设备对后续播放的视频进行画面降质,减少需要进行传输的视频数据,从而在视频会议过程中较好地改善网络传输资源的占用情况,提高网络传输资源的利用率。
进一步地,相比于相关技术中在第二设备视频画面中检测感兴趣区域和降低非感兴趣区域画质的方法,上述视频播放方法无需对视频画面进行太多的分析和处理,可以节约数据处理资源。
进一步地,上述视频播放方法可以在用户临时离开时进行画面降质,减少数据传输带宽,在检测到用户返回时恢复画质,从而减少用户对于画质的手动操作,提升用户体验,而且,可以在第一设备侧用户未察觉的情况下,减少数据传输带宽,提高网络传输资源的利用率。
应该理解的是,虽然图2、图7和图12的流程图中的各个步骤按照箭头的指示依次显示,但是这些步骤并不是必然按照箭头指示的顺序依次执行。除非本文中有明确的说明,这些步骤的执行并没有严格的顺序限制,这些步骤可以以其它的顺序执行。而且,图2、图7和图12中的至少一部分步骤可以包括多个子步骤或者多个阶段,这些子步骤或者阶段并不必然是在同一时刻执行完成,而是可以在不同的时刻执行,这些子步骤或者阶段的执行顺序也不必然是依次进行,而是可以与其它步骤或者其它步骤的子步骤或者阶段的至少一部分轮流或者交替地执行。
实施例五
本公开实施例五提供了一种视频播放系统,可以包括:第一设备和第二设备;
所述第二设备,设置为发送第一原始视频画面至所述第一设备;
所述第一设备,设置为在播放所述第一原始视频画面时,采集本地传感数据,根据所述本地传感数据,确定所述第一设备处是否无人,当确定所述第一设备处无人,生成无人观看通知,并发送所述无人观看通知至所述第二设备;
所述第二设备,还设置为在接收到所述无人观看通知时,对第二原始视频画面进行画面降质,得到降质视频画面,并向所述第一设备返回所述降质视频画面,其中,所述画面降质包括降低分辨率和/或降低帧率;
所述第一设备,还设置为播放所述降质视频画面。
可选地,所述第一设备,还设置为当在所述第一设备上播放所述降质视频画面时,若根据所述本地传感数据,确定所述第一设备处有人,则生成有人观看通知,并发送所述有人观看通知至所述第二设备;
所述第二设备,还设置为在接收到所述有人观看通知时,停止对第三原始视频画面进行画面降质,并发送所述第三原始视频画面至所述第一设备;
所述第一设备,还设置为在所述第一设备上播放所述第三原始视频画面。
实施例六
图13是本公开实施例六提供的一种视频播放装置的结构示意图。参考图13,本实施例提供的视频播放装置1300可以包括:采集模块1301、确定模块1302、通知生成模块1303、通知发送模块1304和播放模块1305。其中:
采集模块1301,设置为当第一设备播放第一原始视频画面时,采集本地传感数据;所述第一原始视频画面由第二设备在当前时刻采集得到;
确定模块1302,设置为根据所述本地传感数据,确定第一设备处是否无人;
通知生成模块1303,设置为若所述第一设备处无人,则生成无人观看通知;
通知发送模块1304,设置为发送所述无人观看通知至所述第二设备,以供所述第二设备在接收到所述无人观看通知时,对第二原始视频画面进行画面降质,得到降质视频画面,并发送所述降质视频画面至所述第一设备;所述第二原始视频画面由所述第二设备在下一时刻采集得到,其中,所述画面降质包括降低分辨率和/或降低帧率;
播放模块1305,设置为播放所述降质视频画面。
此处需要说明的是,上述采集模块1301、确定模块1302、通知生成模块1303、通知发送模块1304和播放模块1305可以作为装置的一部分运行在终端中,可以通过终端中的处理器来执行上述模块实现的功能,终端也可以是智能手机(如Android手机、iOS手机等)、平板电脑、掌声电脑以及移动互联网设备(Mobile Internet Devices,MID)、PAD等终端设备。
本公开实施例提供的技术方案,通过当第一设备播放第一原始视频画面时,采集本地传感数据,根据本地传感数据确定第一设备处是否无人,可以检测第一设备处是否无人观看第一原始视频画面,若第一设备处无人,则生成无人观看通知,发送无人观看通知至第二设备,可以在第一设备处无人观看第一原始视频画面时通知第二设备,第二设备在接收到通知时对第二原始视频画面进行画面降质,得到降质视频画面,并发送降质视频画面至第一设备,第一设备播放降质视频画面,可以在第一设备处无人观看时,通知第二设备对后续播放的视频进行画面降质,减少需要进行传输的视频数据,从而减少对网络传输资源的占用,避免视频播放出现缓慢和卡顿的现象。
在上述实施例的基础上,所述本地传感数据包括所述第一设备的图像采集设备采集的图像帧数据。
在另一个实施例中,所述确定模块1302,还可以设置为:依次获取图像帧数据,并对所述图像帧数据进行人脸检测;当在所述图像帧数据中检测到人脸图像时,确定所述第一设备处有人;当确认连续预设次数在获取的所述图像帧数据中未检测到人脸,确定所述第一设备处无人。
在另一个实施例中,所述确定模块1302,还可以设置为:统计所述音频数据中的静默时长,其中,所述静默时长为所述音频数据中未检测到有人发声的时长;将所述静默时长与预设的时长阈值相比较;当所述静默时长未超过所述时长阈值时,确定所述第一设备处有人;当所述静默时长超过所述时长阈值时,确定所述第一设备处无人。
在另一个实施例中,所述确定模块1302,还可以设置为:判断所述红外数据是否处于预设的红外数据区间内;当所述红外数据处于所述红外数据区间内时,确定所述第一设备处有人;当所述红外数据未处于所述红外数据区间内时,确定所述第一设备处无人。
在上述实施例的基础上,所述第二设备为终端或服务器;当所述第二设备为所述终端时,所述第二原始视频画面为所述第二设备采集到的视频画面;当所述第二设备为所述服务器时,所述第二原始视频画面为由第三设备采集、并发送至所述第二设备的视频画面。
在上述实施例的基础上,还包括:
第二通知生成模块,设置为当在所述第一设备上播放所述降质视频画面时,若根据所述本地传感数据,确定所述第一设备处有人,则生成有人观看通知;
第二通知发送模块,设置为发送所述有人观看通知至所述第二设备,以供所述第二设备在接收到所述有人观看通知时,停止对第三原始视频画面进行画面降质,并向所述第一设备返回所述第三原始视频画面;
第二播放模块,设置为在所述第一设备上播放所述第三原始视频画面。
此处需要说明的是,上述第二通知生成模块、第二通知发送模块和第二播放模块可以作为装置的一部分运行在终端中,可以通过终端中的处理器来执行 上述模块实现的功能。
实施例七
图14是本公开实施例七提供的一种视频播放装置的结构示意图。参考图14,本实施例提供的视频播放装置1400可以包括:第一发送模块1401、画面降质模块1402和第二发送模块1403。其中:
第一发送模块1401,设置为发送第一原始视频画面至第一设备;
画面降质模块1402,设置为当接收到所述第一设备发送的无人观看通知时,对第二原始视频画面进行画面降质,得到降质视频画面,其中,所述无人观看通知为所述第一设备在播放所述第一原始视频画面时,采集本地传感数据,根据所述本地传感数据确定所述第一设备处无人时,所生成的通知,其中,所述画面降质包括降低分辨率和/或降低帧率;
第二发送模块1403,设置为发送所述降质视频画面至所述第一设备。
此处需要说明的是,上述第一发送模块1401、画面降质模块1402和第二发送模块1403可以作为装置的一部分运行在终端中,可以通过终端中的处理器来执行上述模块实现的功能。
在上述实施例的基础上,视频播放装置还包括:
原始视频画面获取模块,设置为采集所述第二原始视频画面;或,接收第三设备所采集的所述第二原始视频画面。
此处需要说明的是,上述原始视频画面获取模块可以作为装置的一部分运行在终端中,可以通过终端中的处理器来执行上述模块实现的功能。
在上述实施例的基础上,视频播放装置还包括:
画面降质停止模块,设置为在发送所述降质视频画面至所述第一设备之后,当接收到所述第一设备发送的有人观看通知时,停止对第三原始视频画面进行画面降质,其中,所述有人观看通知为所述第一设备在播放所述降质视频画面时,根据所述本地传感数据确定所述第一设备处有人时,所生成的通知;
第三发送模块,设置为发送所述第三原始视频画面至所述第一设备。
此处需要说明的是,上述画面降质停止模块和第三发送模块可以作为装置的一部分运行在终端中,可以通过终端中的处理器来执行上述模块实现的功能。
上述提供的视频播放系统、视频播放装置可设置为执行上述任意实施例提供的视频播放方法,具备相应的功能和有益效果。
该实施例关于视频播放系统、视频播放装置的具体限定可以参见上文中对于视频播放方法的限定,在此不再赘述。上述视频播放系统、视频播放装置中的各个模块可全部或部分通过软件、硬件及其组合来实现。上述各模块可以硬件形式内嵌于或独立于计算机设备中的处理器中,也可以以软件形式存储于计算机设备中的存储器中,以便于处理器调用执行以上各个模块对应的操作。
实施例八
图15是本公开实施例八提供的一种电子设备的结构示意图。如图15所示,该电子设备包括:处理器40、存储器41、具有触摸功能的显示屏42、输入装置43、输出装置44以及通信装置45。该电子设备中处理器40的数量可以是一个或者多个,图中以一个处理器40为例。该电子设备中存储器41的数量可以是一个或者多个,图中以一个存储器41为例。该电子设备的处理器40、存储器41、显示屏42、输入装置43、输出装置44以及通信装置45可以通过总线或者其他方式连接,图中以通过总线连接为例。在该实施例中,电子设备可以是电脑、手机、平板、投影仪或交互智能平板等。在该实施例中,以电子设备为交互智能平板为例,进行描述。
存储器41作为一种计算机可读存储介质,可设置为存储软件程序、计算机可执行程序以及模块,如本公开任意实施例所述的基于多导联多结构聚合网络的心电信号识别方法对应的程序指令/模块。存储器41可主要包括存储程序区和存储数据区,其中,存储程序区可存储操作系统、至少一个功能所需的应用程序;存储数据区可存储根据设备的使用所创建的数据等。此外,存储器41可以包括高速随机存取存储器,还可以包括非易失性存储器,例如至少一个磁盘存 储器件、闪存器件、或其他非易失性固态存储器件。在一些实例中,存储器41可进一步包括相对于处理器40远程设置的存储器,这些远程存储器可以通过网络连接至设备。上述网络的实例包括但不限于互联网、企业内部网、局域网、移动通信网及其组合。
显示屏42为具有触摸功能的显示屏42,其可以是电容屏、电磁屏或者红外屏。一般而言,显示屏42设置为根据处理器40的指示显示数据,还设置为接收作设置为显示屏42的触摸操作,并将相应的信号发送至处理器40或其他装置。可选地,当显示屏42为红外屏时,其还包括红外触摸框,该红外触摸框设置在显示屏42的四周,其还可以设置为接收红外信号,并将该红外信号发送至处理器40或者其他设备。
通信装置45,设置为与其他设备建立通信连接,其可以是有线通信装置和/或无线通信装置。
输入装置43可设置为接收输入的数字或者字符信息,以及产生与电子设备的用户设置以及功能控制有关的键信号输入,还可以是设置为获取图像的摄像头以及获取音频数据的拾音设备。输出装置44可以包括扬声器等音频设备。需要说明的是,输入装置43和输出装置44的具体组成可以根据实际情况设定。
处理器40通过运行存储在存储器41中的软件程序、指令以及模块,从而执行设备的各种功能应用以及数据处理,即实现上述的基于多导联多结构聚合网络的心电信号识别方法。
可选地,在该实施例中,处理器40执行存储器41中存储的一个或多个程序时,可以实现如下操作:
当第一设备播放第一原始视频画面时,采集本地传感数据;
根据所述本地传感数据,确定第一设备处是否无人;
当确定所述第一设备处无人,生成无人观看通知;
发送所述无人观看通知至第二设备,以供所述第二设备在接收到所述无人观看通知时,对第二原始视频画面进行画面降质,得到降质视频画面,并向所述第一设备返回所述降质视频画面,其中,所述画面降质包括降低分辨率和/或降低帧率;
播放所述降质视频画面。
在上述实施例的基础上,一个或多个处理器40还可以实现如下操作:
所述本地传感数据包括所述第一设备的图像采集设备采集的图像帧数据。
在上述实施例的基础上,一个或多个处理器40还可以实现如下操作:
依次获取图像帧数据,并对所述图像帧数据进行人脸检测;
当在所述图像帧数据中检测到人脸图像时,确定所述第一设备处有人;
当确认连续预设次数在获取的所述图像帧数据中未检测到人脸,确定所述第一设备处无人。
在上述实施例的基础上,一个或多个处理器40还实现如下操作:
统计所述音频数据中的静默时长,其中,所述静默时长为所述音频数据中未检测到有人发声的时长;
将所述静默时长与预设的时长阈值相比较;
当所述静默时长未超过所述时长阈值时,确定所述第一设备处有人;
当所述静默时长超过所述时长阈值时,确定所述第一设备处无人。
在上述实施例的基础上,一个或多个处理器40还可以实现如下操作:
判断所述红外数据是否处于预设的红外数据区间内;
当所述红外数据处于所述红外数据区间内时,确定所述第一设备处有人;
当所述红外数据未处于所述红外数据区间内时,确定所述第一设备处无人。
在上述实施例的基础上,一个或多个处理器40还可以实现如下操作:
所述第二设备为终端或服务器;当所述第二设备为所述终端时,所述第二原始视频画面为所述第二设备采集到的视频画面;当所述第二设备为所述服务器时,所述第二原始视频画面为由第三设备采集、并发送至所述第二设备的视频画面。
在上述实施例的基础上,一个或多个处理器40还可以实现如下操作:
当在所述第一设备上播放所述降质视频画面时,若根据所述本地传感数据,确定所述第一设备处有人,则生成有人观看通知;
发送所述有人观看通知至所述第二设备,以供所述第二设备在接收到所述有人观看通知时,停止对第三原始视频画面进行画面降质,并向所述第一设备 返回所述第三原始视频画面;
在所述第一设备上播放所述第三原始视频画面。
在上述实施例的基础上,一个或多个处理器40还可以实现如下操作:
发送第一原始视频画面至第一设备;
当接收到所述第一设备发送的无人观看通知时,对第二原始视频画面进行画面降质,得到降质视频画面,其中,所述无人观看通知为所述第一设备在播放所述第一原始视频画面时,采集本地传感数据,根据所述本地传感数据确定所述第一设备处无人时,所生成的通知,其中,所述画面降质包括降低分辨率和/或降低帧率;
发送所述降质视频画面至所述第一设备。
在上述实施例的基础上,一个或多个处理器40还可以实现如下操作:
采集所述第二原始视频画面;
或,
接收第三设备所采集的所述第二原始视频画面。
在上述实施例的基础上,一个或多个处理器40还实现如下操作:
在发送所述降质视频画面至所述第一设备之后,当接收到所述第一设备发送的有人观看通知时,停止对第三原始视频画面进行画面降质,其中,所述有人观看通知为所述第一设备在播放所述降质视频画面时,根据所述本地传感数据确定所述第一设备处有人时,所生成的通知;
发送所述第三原始视频画面至所述第一设备。
实施例九
本公开实施例九还提供一种包含计算机可执行指令的非易失性存储介质。图16是本公开实施例九提供的一种电子设备的结构示意图。如图16所示,描述了根据本申请的实施方式的程序产品160,其上存储有计算机可执行指令,所述计算机可执行指令在由计算机处理器执行时用于执行一种视频播放方法,包括:
当第一设备播放第一原始视频画面时,采集本地传感数据;
根据所述本地传感数据,确定第一设备处是否无人;
当确定所述第一设备处无人,生成无人观看通知;
发送所述无人观看通知至第二设备,以供所述第二设备在接收到所述无人观看通知时,对第二原始视频画面进行画面降质,得到降质视频画面,并向所述第一设备返回所述降质视频画面,其中,所述画面降质包括降低分辨率和/或降低帧率;
播放所述降质视频画面。
可选地,所述计算机可执行指令在由计算机处理器执行时用于执行如下步骤的程序代码:
所述本地传感数据包括所述第一设备的图像采集设备采集的图像帧数据。
可选地,所述计算机可执行指令在由计算机处理器执行时用于执行如下步骤的程序代码:
依次获取图像帧数据,并对所述图像帧数据进行人脸检测;
当在所述图像帧数据中检测到人脸图像时,确定所述第一设备处有人;
当确认连续预设次数在获取的所述图像帧数据中未检测到人脸,确定所述第一设备处无人。
可选地,所述计算机可执行指令在由计算机处理器执行时用于执行如下步骤的程序代码:
统计所述音频数据中的静默时长,其中,所述静默时长为所述音频数据中未检测到有人发声的时长;
将所述静默时长与预设的时长阈值相比较;
当所述静默时长未超过所述时长阈值时,确定所述第一设备处有人;
当所述静默时长超过所述时长阈值时,确定所述第一设备处无人。
可选地,所述计算机可执行指令在由计算机处理器执行时用于执行如下步骤的程序代码:
判断所述红外数据是否处于预设的红外数据区间内;
当所述红外数据处于所述红外数据区间内时,确定所述第一设备处有人;
当所述红外数据未处于所述红外数据区间内时,确定所述第一设备处无人。
可选地,所述计算机可执行指令在由计算机处理器执行时用于执行如下步骤的程序代码:
所述第二设备为终端或服务器;当所述第二设备为所述终端时,所述第二原始视频画面为所述第二设备采集到的视频画面;当所述第二设备为所述服务器时,所述第二原始视频画面为由第三设备采集、并发送至所述第二设备的视频画面。
可选地,所述计算机可执行指令在由计算机处理器执行时用于执行如下步骤的程序代码:
当在所述第一设备上播放所述降质视频画面时,若根据所述本地传感数据,确定所述第一设备处有人,则生成有人观看通知;
发送所述有人观看通知至所述第二设备,以供所述第二设备在接收到所述有人观看通知时,停止对第三原始视频画面进行画面降质,并向所述第一设备返回所述第三原始视频画面;
在所述第一设备上播放所述第三原始视频画面。
可选地,所述计算机可执行指令在由计算机处理器执行时用于执行一种视频播放方法,包括:
发送第一原始视频画面至第一设备;
当接收到所述第一设备发送的无人观看通知时,对第二原始视频画面进行画面降质,得到降质视频画面,其中,所述无人观看通知为所述第一设备在播放所述第一原始视频画面时,采集本地传感数据,根据所述本地传感数据确定所述第一设备处无人时,所生成的通知,其中,所述画面降质包括降低分辨率和/或降低帧率;
发送所述降质视频画面至所述第一设备。
可选地,所述计算机可执行指令在由计算机处理器执行时用于执行如下步骤的程序代码:
采集所述第二原始视频画面;
或,
接收第三设备所采集的所述第二原始视频画面。
可选地,所述计算机可执行指令在由计算机处理器执行时用于执行如下步骤的程序代码:
在发送所述降质视频画面至所述第一设备之后,当接收到所述第一设备发送的有人观看通知时,停止对第三原始视频画面进行画面降质,其中,所述有人观看通知为所述第一设备在播放所述降质视频画面时,根据所述本地传感数据确定所述第一设备处有人时,所生成的通知;
发送所述第三原始视频画面至所述第一设备。
当然,本公开实施例所提供的一种包含计算机可执行指令的非易失性存储介质,其计算机可执行指令不限于如上所述的视频播放方法的操作,还可以执行本公开任意实施例所提供的视频播放方法中的相关操作,且具备相应的功能和有益效果。
本领域普通技术人员可以理解实现上述实施例方法中的全部或部分流程,是可以通过计算机程序来指令相关的硬件来完成,所述的计算机程序可存储于一非易失性计算机可读取存储介质中,该计算机程序在执行时,可包括如上述各方法的实施例的流程。其中,本公开所提供的各实施例中所使用的对存储器、存储、数据库或其它介质的任何引用,均可包括非易失性和/或易失性存储器。非易失性存储器可包括只读存储器(ROM)、可编程ROM(PROM)、电可编程ROM(EPROM)、电可擦除可编程ROM(EEPROM)或闪存。易失性存储器可包括随机存取存储器(RAM)或者外部高速缓冲存储器。作为说明而非局限,RAM以多种形式可得,诸如静态RAM(SRAM)、动态RAM(DRAM)、同步DRAM(SDRAM)、双数据率SDRAM(DDRSDRAM)、增强型SDRAM(ESDRAM)、同步链路(Synchlink)DRAM(SLDRAM)、存储器总线(Rambus)直接RAM(RDRAM)、直接存储器总线动态RAM(DRDRAM)、以及存储器总线动态RAM(RDRAM)等。
以上实施例的各技术特征可以进行任意的组合,为使描述简洁,未对上述实施例中的各个技术特征所有可能的组合都进行描述,然而,只要这些技术特 征的组合不存在矛盾,都应当认为是本说明书记载的范围。
以上所述实施例仅表达了本公开的几种实施方式,其描述较为具体和详细,但并不能因此而理解为对发明专利范围的限制。应当指出的是,对于本领域的普通技术人员来说,在不脱离本公开构思的前提下,还可以做出若干变形和改进,这些都属于本公开的保护范围。因此,本公开专利的保护范围应以所附权利要求为准。

Claims (14)

  1. 一种视频播放方法,包括:
    当第一设备播放第一原始视频画面时,采集本地传感数据;
    根据所述本地传感数据,确定所述第一设备处是否无人;
    当确定所述第一设备处无人时,生成无人观看通知;
    发送所述无人观看通知至第二设备,以供所述第二设备在接收到所述无人观看通知时,对第二原始视频画面进行画面降质,得到降质视频画面,并向所述第一设备返回所述降质视频画面,其中,所述画面降质包括降低分辨率和/或降低帧率;
    在所述第一设备上播放所述降质视频画面。
  2. 根据权利要求1所述的方法,其中,所述本地传感数据包括所述第一设备的图像采集设备采集的图像帧数据。
  3. 根据权利要求2所述的方法,其中,所述根据所述本地传感数据,确定所述第一设备处是否无人,包括:
    依次获取所述图像帧数据,并对所述图像帧数据进行人脸检测;
    当在所述图像帧数据中检测到人脸图像时,确定所述第一设备处有人;
    当连续预设次数在获取的所述图像帧数据中未检测到所述人脸图像时,确定所述第一设备处无人。
  4. 根据权利要求1所述的方法,其中,所述本地传感数据还包括所述第一设备的音频采集设备采集的音频数据;所述根据所述本地传感数据,确定所述第一设备处是否无人,包括:
    统计所述音频数据中的静默时长,其中,所述静默时长为所述音频数据中未检测到有人发声的时长;
    将所述静默时长与预设的时长阈值相比较;
    当所述静默时长未超过所述时长阈值时,确定所述第一设备处有人;
    当所述静默时长超过所述时长阈值时,确定所述第一设备处无人。
  5. 根据权利要求1所述的方法,其中,所述本地传感数据还包括所述第一设备的红外传感设备采集的红外数据,所述根据所述本地传感数据,确定所述第一设备处是否无人,包括:
    判断所述红外数据是否处于预设的红外数据区间内;
    当所述红外数据处于所述红外数据区间内时,确定所述第一设备处有人;
    当所述红外数据未处于所述红外数据区间内时,确定所述第一设备处无人。
  6. 根据权利要求1所述的视频播放方法,其中,所述第二设备为终端或服务器;当所述第二设备为所述终端时,所述第二原始视频画面为所述第二设备采集到的视频画面;当所述第二设备为所述服务器时,所述第二原始视频画面为由第三设备采集、并发送至所述第二设备的视频画面。
  7. 根据权利要求1所述的视频播放方法,其中,当在所述第一设备上播放所述降质视频画面时,所述方法还包括:
    若根据所述本地传感数据,确定所述第一设备处有人,则生成有人观看通知;
    发送所述有人观看通知至所述第二设备,以供所述第二设备在接收到所述有人观看通知时,停止对第三原始视频画面进行画面降质,并向所述第一设备返回所述第三原始视频画面;
    在所述第一设备上播放所述第三原始视频画面。
  8. 一种视频播放方法,包括:
    发送第一原始视频画面至第一设备;
    当接收到所述第一设备发送的无人观看通知时,对第二原始视频画面进行画面降质,得到降质视频画面,其中,所述无人观看通知为所述第一设备在播放所述第一原始视频画面时,采集本地传感数据,当根据所述本地传感数据确定所述第一设备处无人时,所生成的通知,所述画面降质包括降低分辨率和/或降低帧率;
    发送所述降质视频画面至所述第一设备。
  9. 根据权利要求8所述的视频播放方法,其中,所述方法还包括:
    采集所述第二原始视频画面;
    或,
    接收第三设备所采集的所述第二原始视频画面。
  10. 根据权利要求8所述的视频播放方法,其中,在发送所述降质视频画面至所述第一设备之后,所述方法还包括:
    当接收到所述第一设备发送的有人观看通知时,停止对第三原始视频画面进行画面降质,其中,所述有人观看通知为所述第一设备在播放所述降质 视频画面时,根据所述本地传感数据确定所述第一设备处有人时,所生成的通知;
    发送所述第三原始视频画面至所述第一设备。
  11. 一种视频播放系统,所述系统包括第一设备和第二设备;
    所述第二设备,设置为发送第一原始视频画面至所述第一设备;
    所述第一设备,设置为在播放所述第一原始视频画面时,采集本地传感数据,根据所述本地传感数据,确定所述第一设备处是否无人,当确定所述第一设备处无人时,生成无人观看通知,并发送所述无人观看通知至所述第二设备;
    所述第二设备,还设置为在接收到所述无人观看通知时,对第二原始视频画面进行画面降质,得到降质视频画面,并向所述第一设备返回所述降质视频画面,所述画面降质包括降低分辨率和/或降低帧率;
    所述第一设备,还设置为播放所述降质视频画面。
  12. 根据权利要求11所述的视频播放系统,其中,所述第一设备,还设置为当在所述第一设备上播放所述降质视频画面时,若根据所述本地传感数据,确定所述第一设备处有人,则生成有人观看通知,并发送所述有人观看通知至所述第二设备;
    所述第二设备,还设置为在接收到所述有人观看通知时,停止对第三原始视频画面进行画面降质,并发送所述第三原始视频画面至所述第一设备;
    所述第一设备,还设置为播放所述第三原始视频画面。
  13. 一种电子设备,包括:存储器、一个或多个处理器;
    所述存储器,设置为存储一个或多个程序;
    当所述一个或多个程序被所述一个或多个处理器执行,使得所述一个或多个处理器执行如权利要求1-10中任一所述的视频播放方法。
  14. 一种包含计算机可执行指令的非易失性存储介质,所述计算机可执行指令在由计算机处理器执行时用于执行如权利要求1-10中任一所述的视频播放方法。
PCT/CN2020/138146 2020-11-02 2020-12-21 视频播放方法、系统、电子设备和存储介质 WO2022088447A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202011202335.3 2020-11-02
CN202011202335.3A CN112584083B (zh) 2020-11-02 2020-11-02 视频播放方法、系统、电子设备和存储介质

Publications (1)

Publication Number Publication Date
WO2022088447A1 true WO2022088447A1 (zh) 2022-05-05

Family

ID=75119981

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/138146 WO2022088447A1 (zh) 2020-11-02 2020-12-21 视频播放方法、系统、电子设备和存储介质

Country Status (2)

Country Link
CN (1) CN112584083B (zh)
WO (1) WO2022088447A1 (zh)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117440209A (zh) * 2023-12-15 2024-01-23 牡丹江师范学院 一种基于演唱场景的实现方法及系统

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113347461B (zh) * 2021-07-05 2022-11-11 湖南快乐阳光互动娱乐传媒有限公司 一种媒资内容提升方法及装置
CN114827664B (zh) * 2022-04-27 2023-10-20 咪咕文化科技有限公司 多路直播混流方法、服务器、终端设备、系统及存储介质

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130021535A1 (en) * 2011-07-18 2013-01-24 Lg Electronics Inc. Electronic device and operating method thereof
CN103347165A (zh) * 2013-07-25 2013-10-09 广东威创视讯科技股份有限公司 一种视频数据处理方法及终端
CN106686257A (zh) * 2017-03-10 2017-05-17 广东欧珀移动通信有限公司 一种移动终端的控制方法、装置及移动终端
CN107404670A (zh) * 2016-05-18 2017-11-28 中国移动通信集团北京有限公司 一种视频播放控制方法及装置
CN108495116A (zh) * 2018-03-29 2018-09-04 京东方科技集团股份有限公司 3d显示装置及其控制方法、计算机设备
CN108881781A (zh) * 2018-07-17 2018-11-23 广东小天才科技有限公司 视频通话过程中分辨率的确定方法及装置
CN108966010A (zh) * 2018-07-10 2018-12-07 南阳师范学院 在线视频分辨率的调节方法、装置、存储介质及终端
CN111510785A (zh) * 2020-04-16 2020-08-07 Oppo广东移动通信有限公司 视频播放的控制方法、装置、终端和计算机可读存储介质

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
BR102015005929B1 (pt) * 2015-03-17 2021-11-30 Samsung Eletrônica Da Amazônia Ltda Sistema e método para composição e compressão de vídeo com base em contexto a partir de objetos com resolução espacial normalizada
CN107317992A (zh) * 2017-07-19 2017-11-03 苏州三星电子电脑有限公司 视频通信控制装置与控制方法
CN108377358B (zh) * 2018-03-31 2020-07-07 江苏怀业安全科技有限公司 视频通信处理方法、装置及计算机存储介质
CN110572607A (zh) * 2019-08-20 2019-12-13 视联动力信息技术股份有限公司 一种视频会议方法、系统及装置和存储介质
CN110557597A (zh) * 2019-09-12 2019-12-10 视联动力信息技术股份有限公司 一种视频会议签到方法、服务器、电子设备及存储介质

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130021535A1 (en) * 2011-07-18 2013-01-24 Lg Electronics Inc. Electronic device and operating method thereof
CN103347165A (zh) * 2013-07-25 2013-10-09 广东威创视讯科技股份有限公司 一种视频数据处理方法及终端
CN107404670A (zh) * 2016-05-18 2017-11-28 中国移动通信集团北京有限公司 一种视频播放控制方法及装置
CN106686257A (zh) * 2017-03-10 2017-05-17 广东欧珀移动通信有限公司 一种移动终端的控制方法、装置及移动终端
CN108495116A (zh) * 2018-03-29 2018-09-04 京东方科技集团股份有限公司 3d显示装置及其控制方法、计算机设备
CN108966010A (zh) * 2018-07-10 2018-12-07 南阳师范学院 在线视频分辨率的调节方法、装置、存储介质及终端
CN108881781A (zh) * 2018-07-17 2018-11-23 广东小天才科技有限公司 视频通话过程中分辨率的确定方法及装置
CN111510785A (zh) * 2020-04-16 2020-08-07 Oppo广东移动通信有限公司 视频播放的控制方法、装置、终端和计算机可读存储介质

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117440209A (zh) * 2023-12-15 2024-01-23 牡丹江师范学院 一种基于演唱场景的实现方法及系统
CN117440209B (zh) * 2023-12-15 2024-03-01 牡丹江师范学院 一种基于演唱场景的实现方法及系统

Also Published As

Publication number Publication date
CN112584083B (zh) 2022-05-27
CN112584083A (zh) 2021-03-30

Similar Documents

Publication Publication Date Title
WO2022088447A1 (zh) 视频播放方法、系统、电子设备和存储介质
US9788017B2 (en) Video analytics with pre-processing at the source end
US9667908B2 (en) Image recording system
US20140146877A1 (en) Method for dynamically adapting video image parameters for facilitating subsequent applications
JP2009510877A (ja) 顔検出を利用したストリーミングビデオにおける顔アノテーション
CN110753202B (zh) 可视对讲系统的音视频同步方法、装置、设备及存储介质
TW201503054A (zh) 影像處理器及其影像合成方法
WO2019149066A1 (zh) 视频播放方法、终端设备及存储介质
KR20110052933A (ko) 촬영장치 및 촬영영상 제공방법
US10798366B2 (en) Motion detection device and motion detection method
WO2022048129A1 (zh) 对象识别方法、装置及系统
CN112235600B (zh) 视频数据及视频业务请求的处理方法、装置及系统
US11095901B2 (en) Object manipulation video conference compression
US11706389B2 (en) Data transmission method, camera and electronic device
US10762913B2 (en) Image-based techniques for audio content
WO2019076156A1 (zh) 数据传输方法及摄像机、电子设备
US10846810B2 (en) Leveraging audio/video recording and communication devices to enhance public safety
US9830946B2 (en) Source data adaptation and rendering
CN101420592A (zh) 具有保安功能的四画面分割对讲系统
CN113316041B (zh) 远程健康检测系统、方法、装置及设备
US20230276081A1 (en) System and method of altering real time audio and video streams
CN112272284B (zh) 基于双摄像头的多方视频通信方法、装置和系统
US20220286621A1 (en) Event video sequences
US20240062580A1 (en) Visual tracking system for active object
CN114125465A (zh) 录像的存储方法、装置、系统、存储介质以及计算机设备

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20959604

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20959604

Country of ref document: EP

Kind code of ref document: A1