CN111988555B - Data processing method, device, equipment and machine readable medium - Google Patents

Data processing method, device, equipment and machine readable medium Download PDF

Info

Publication number
CN111988555B
CN111988555B CN201910426233.0A CN201910426233A CN111988555B CN 111988555 B CN111988555 B CN 111988555B CN 201910426233 A CN201910426233 A CN 201910426233A CN 111988555 B CN111988555 B CN 111988555B
Authority
CN
China
Prior art keywords
video
conference
operation interface
initiating
user
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910426233.0A
Other languages
Chinese (zh)
Other versions
CN111988555A (en
Inventor
管赛南
姚维
沈杰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Banma Zhixing Network Hongkong Co Ltd
Original Assignee
Banma Zhixing Network Hongkong Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Banma Zhixing Network Hongkong Co Ltd filed Critical Banma Zhixing Network Hongkong Co Ltd
Priority to CN201910426233.0A priority Critical patent/CN111988555B/en
Publication of CN111988555A publication Critical patent/CN111988555A/en
Application granted granted Critical
Publication of CN111988555B publication Critical patent/CN111988555B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/15Conference systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/478Supplemental services, e.g. displaying phone caller identification, shopping application
    • H04N21/4788Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting

Abstract

The embodiment of the application provides a data processing method, a data processing device, equipment and a machine readable medium, wherein the method comprises the following steps: collecting a video corresponding to a conference environment; displaying a video corresponding to the conference environment in an operation interface; wherein, the operation interface includes: receiving an operation interface of the video conference invitation; or an operation interface for initiating a video conference. According to the video conference method and device, the user can be prompted to use the operation function corresponding to the video in the operation interface, namely the video conference can be unfolded, and the user can also be prompted to operate the operation interface, so that the probability of misoperation can be reduced.

Description

Data processing method, device, equipment and machine readable medium
Technical Field
The present application relates to the field of communications technologies, and in particular, to a data processing method, a data processing apparatus, a device, and a machine-readable medium.
Background
The video conference system is an interactive multimedia information service integrating voice, image and data, and is a value-added service based on a communication network. In a conference carried out by adopting the video conference system, the participating users are distributed in different region spaces, and the real-time mutual communication and transmission of sound information and image information of the participating users can be realized through the video conference system. A conference terminal is a device used by a participating user to participate in a conference. The conference terminal may function as a conference operation function, for example, the conference operation function may include: initiating a video conference, receiving a conference invitation and the like.
The conference terminal can provide a conference operation interface so that a user can realize a conference operation function through the conference operation interface. The current conference operation interface may include: and prompting information such as a text and an operation control so that a user can acquire a conference operation function corresponding to the conference operation interface through the prompting text, and further can execute corresponding operation through the operation control. Taking the conference operation interface as the conference invitation interface as an example, the conference invitation interface may include: prompt text "meeting invitation from user a", "answer control", and "decline control", the prompt text being available for the user to select one of the "answer control" and "decline control" to perform an operation.
In practical application, the difficulty of the user to acquire the conference operation function through the prompt text is easily increased due to the factors such as inattention and the like, so that the user can make wrong judgment on the function and execute corresponding misoperation. For example, the conference operation function corresponding to the conference operation interface is "receive video conference invitation", but the user determines that the conference call invitation is received by mistake, and therefore a corresponding misoperation is performed on the "reject control".
Disclosure of Invention
The technical problem to be solved by the embodiments of the present application is to provide a data processing method, which can prompt a user to use an operation function corresponding to a video in an operation interface to expand a video conference, and can also prompt the user to operate the operation interface, so that the probability of misoperation can be reduced.
Correspondingly, the embodiment of the application also provides a data processing device, equipment and a machine readable medium, which are used for ensuring the realization and the application of the method.
In order to solve the above problem, an embodiment of the present application discloses a data processing method, including:
collecting a video corresponding to a conference environment;
displaying a video corresponding to the conference environment in an operation interface;
wherein, the operation interface includes:
receiving an operation interface of the video conference invitation; or
And initiating an operation interface of the video conference.
In order to solve the above problem, an embodiment of the present application discloses a data processing method, including:
collecting a video corresponding to a conference environment;
displaying the video as a background of an operation interface;
wherein, the operation interface includes:
receiving an operation interface of the video conference invitation; or
And initiating an operation interface of the video conference.
On the other hand, the embodiment of the present application further discloses a data processing apparatus, including:
the video acquisition module is used for determining a video corresponding to the conference environment; and
the video display module is used for displaying the video corresponding to the conference environment in an operation interface;
wherein, the operation interface includes:
receiving an operation interface of the video conference invitation; or
And initiating an operation interface of the video conference.
On the other hand, the embodiment of the present application further discloses a data processing apparatus, including:
the video acquisition module is used for determining a video corresponding to the conference environment; and
the video display module is used for displaying the video as the background of the operation interface;
wherein, the operation interface includes:
receiving an operation interface of the video conference invitation; or
And initiating an operation interface of the video conference.
In another aspect, an embodiment of the present application further discloses an apparatus, including:
one or more processors; and
one or more machine-readable media having instructions stored thereon that, when executed by the one or more processors, cause the apparatus to perform one or more of the methods described above.
In yet another aspect, embodiments of the present application disclose one or more machine-readable media having instructions stored thereon, which when executed by one or more processors, cause an apparatus to perform one or more of the methods described above.
Compared with the background art, the embodiment of the application has the following advantages:
according to the video conference system and the video conference method, the video corresponding to the conference environment is displayed in the operation interface, and the video attribute of the operation interface can be increased through the video in the operation interface, so that more bases can be provided for understanding of the interface operation function, and specifically, a user can be prompted to use the operation function corresponding to the video in the operation interface to expand the video conference.
Moreover, the video attribute of the operation interface is added to the video in the operation interface, more bases can be provided for the operation corresponding to the operation interface, and the user can be prompted to operate the operation interface, so that the probability of misoperation can be reduced.
Drawings
Fig. 1 is a schematic structural diagram of a video conference system according to an embodiment of the present application;
FIG. 2 is a flow chart of steps of a first embodiment of a data processing method of the present application;
FIG. 3 is an illustration of an interface for receiving a video conference invitation according to an embodiment of the disclosure;
FIG. 4 is a schematic view of an operation interface for initiating a video conference according to an embodiment of the present application;
FIG. 5 is a flowchart illustrating steps of a second embodiment of a data processing method according to the present application;
fig. 6 is a schematic structural diagram of a video conference system according to an embodiment of the present application;
FIG. 7 is a flowchart illustrating the steps of a third embodiment of a data processing method according to the present application;
FIG. 8 is a block diagram of an embodiment of a data processing apparatus of the present application;
FIG. 9 is a block diagram of a data processing apparatus according to an embodiment of the present application; and
fig. 10 is a schematic structural diagram of an apparatus provided in an embodiment of the present application.
Detailed Description
In order to make the aforementioned objects, features and advantages of the present application more comprehensible, the present application is described in further detail with reference to the accompanying drawings and the detailed description.
The technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are only a part of the embodiments of the present application, and not all of the embodiments. All other embodiments that can be derived from the embodiments given herein by a person of ordinary skill in the art are intended to be within the scope of the present disclosure.
While the concepts of the present application are susceptible to various modifications and alternative forms, specific embodiments thereof have been shown by way of example in the drawings and will herein be described in detail. It should be understood, however, that the description above is not intended to limit the application to the particular forms disclosed, but on the contrary, the intention is to cover all modifications, equivalents, and alternatives falling within the spirit and scope of the application.
Reference in the specification to "one embodiment," "an embodiment," "a particular embodiment," or the like, means that the embodiment described may include a particular feature, structure, or characteristic, but every embodiment may or may not necessarily include the particular feature, structure, or characteristic. Moreover, such phrases are not necessarily referring to the same embodiment. Further, where a particular feature, structure, or characteristic is described in connection with an embodiment, it is submitted that it is within the knowledge of one skilled in the art to affect such feature, structure, or characteristic in connection with other embodiments whether or not explicitly described. In addition, it should be understood that items in the list included in the form "at least one of a, B, and C" may include the following possible items: (A) (ii) a (B) (ii) a (C) (ii) a (A and B); (A and C); (B and C); or (A, B and C). Likewise, a listing of items in the form of "at least one of a, B, or C" may mean (a); (B) (ii) a (C) (ii) a (A and B); (A and C); (B and C); or (A, B and C).
In some cases, the disclosed embodiments may be implemented as hardware, firmware, software, or any combination thereof. The disclosed embodiments may also be implemented as instructions carried or stored on one or more transitory or non-transitory machine-readable (e.g., computer-readable) storage media, which may be executed by one or more processors. A machine-readable storage medium may be implemented as a storage device, mechanism, or other physical structure (e.g., a volatile or non-volatile memory, a media disk, or other media other physical structure device) for storing or transmitting information in a form readable by a machine.
In the drawings, some structural or methodical features may be shown in a particular arrangement and/or ordering. Preferably, however, such specific arrangement and/or ordering is not necessary. Rather, in some embodiments, such features may be arranged in different ways and/or orders than as shown in the figures. Moreover, the inclusion of structural or methodical features in particular figures is not meant to imply that such features are required in all embodiments and that, in some embodiments, such features may not be included or may be combined with other features.
The embodiment of the application can be applied to video conference scenes. Referring to fig. 1, a schematic structural diagram of a video conference system according to an embodiment of the present application is shown, which may specifically include: at least two participating devices 101 and at least one controlling device 102.
The participating devices 101 are devices used by participating users, and a system generally comprises a plurality of participating devices 101. The conferencing device 101 generally includes: a codec, an input device, and an output device. The codec is responsible for preprocessing, encoding, decoding, post-processing, network communication, user control and the like of signals such as voice, video and the like. The input device may include a microphone, a camera, and other acquisition devices, and the output device may include a sound, a television, a display device, and other devices. The conference participating equipment 101 collects the voice, video and other data of the conference participating users, and the data is preprocessed, compressed and encoded, encapsulated into a data packet format and sent to the far end by the network; and meanwhile, a data packet at a far end is received from the network, and effective data obtained after the data packet is unpacked is decoded and then played to a user.
The control device 102 is used for multiparty communication during a conference. During the conference, a many-to-one connection is established between the plurality of participant devices 101 and the control device 102 participating in the multi-party conference communication, and audio data and video data are exchanged between the participant devices 101 through the control device 102. The control device 102 is responsible for completing the switching and mixing of the media streams. For the audio media streams, the control device 102 generally outputs a mixed audio media stream for each terminal, and the mixing generally selects several audio media streams with the largest input volume for superposition. For video streaming, the control device 102 may send a single-picture video stream of one participating device 101 to another participating device 101, or if the control device 102 supports multi-picture functionality, may be able to combine videos from multiple participating devices 101 into one multi-picture image and then send it to a terminal or terminals.
Optionally, the system may further include: and calling the controller. The call controller is used to complete the routing of the call. For example, Gatekeeper (Gatekeeper) entities defined in the h.323 standard and Proxy server (Proxy) entities defined in the SIP (Session Initiation Protocol) standard are responsible for completing call routing functions.
Optionally, the system may further include: and the gateway equipment is used for realizing the conversion of equipment intercommunication among different network protocols and media formats.
The information content exchanged between the participating devices 101 may include: call control signaling and one or more of an audio stream, a video stream, a text message stream, etc. The meaning of the media stream, and the codec and delivery management of the media stream, may be referred to the relevant standards, or may be determined by those skilled in the art according to the actual application requirements.
In practical applications, the participating device 101 may include: and (4) a conference terminal. A conference terminal is a device used by a participating user to participate in a conference. The conference terminal may function as a conference operation function, for example, the conference operation function may include: initiating a video conference, receiving a conference invitation and the like.
The conference terminal may specifically include but is not limited to: smart phones, tablet computers, electronic book readers, MP3 (Moving Picture Experts Group Audio Layer III) players, MP4 (Moving Picture Experts Group Audio Layer IV) players, laptop portable computers, PCs (Personal computers), set-top boxes, smart televisions, wearable devices, smart microphones, and the like. The conference terminal of the embodiment of the application can be arranged in a conference environment, such as an intelligent microphone arranged in the conference environment; the conference terminal of the embodiment of the application can also be independent of a conference environment, such as user equipment including a mobile phone. It is understood that the embodiment of the present application does not impose a limitation on a specific conference terminal.
The conference environment of the embodiment of the application can be used for representing the environment corresponding to the conference, and the conference environment can include: conference sites, etc., which may include: conference rooms, etc.
Optionally, the conference terminal may be internally or externally provided with the codec, the input device, and the output device.
The embodiment of the present application provides a data processing scheme applied to a conference terminal, which is directed to a problem that a user knows a conference operation function through a prompt text with a high difficulty, and the scheme may specifically include: collecting a video corresponding to a conference environment; and displaying the video corresponding to the conference environment in an operation interface.
An Interface (UI) may refer to the overall design of human-computer interaction, operating logic, and an aesthetically pleasing Interface to software. An interface is a medium for interaction and information exchange between a computer system and a user that enables conversion between an internal form of information and a human-acceptable form. The interface is designed between the user and the hardware for interactive communication with the related software, and aims to enable the user to conveniently and efficiently operate the hardware to achieve bidirectional interaction and complete the work which is expected to be completed by the hardware.
The operation interface of the embodiment of the application may specifically include:
receiving an operation interface of the video conference invitation; or
And initiating an operation interface of the video conference.
The video corresponding to the conference environment is displayed in the operation interface, and the video attribute of the interface can be increased through the video in the operation interface, so that more bases can be provided for understanding of the interface function, and specifically, a user can be prompted to use the function corresponding to the video in the operation interface to expand the video conference.
In addition, as the video attribute of the interface is added to the video in the operation interface, more bases can be provided for the operation corresponding to the operation interface, and the user can be prompted to operate the operation interface, so that the probability of misoperation can be reduced.
Method embodiment one
Referring to fig. 2, a flowchart illustrating steps of a first embodiment of a data processing method according to the present application is shown, where the method may specifically include the following steps:
step 201, collecting a video corresponding to a conference environment;
step 202, displaying a video corresponding to the conference environment in an operation interface;
the operation interface may specifically include:
receiving an operation interface of the video conference invitation; or
And initiating an operation interface of the video conference.
In step 201, the video corresponding to the conference environment can be used to describe the conditions of the conference environment, such as the personnel conditions in the conference environment.
In practical application, the terminal may establish a connection with the acquisition device corresponding to the conference environment, so as to receive the video acquired by the acquisition device through the connection.
According to an embodiment, the video data collected by the collecting device can be directly used as the video corresponding to the conference environment.
According to another embodiment, the video data acquired by the acquisition device may be subjected to video processing to obtain a video corresponding to the conference environment. The video processing may include: transcoding processing, resolution processing, beautifying processing, watermarking processing, filter processing and the like. The transcoding process may decode the video received from the capture device (in a first encoded format) and then re-encode it into a second encoded format.
In the embodiment of the present application, the video may include: real-time video or non-real-time video. Wherein the real-time video may describe the real-time situation of the conference environment. Non-real-time video may also characterize the video properties of the operator interface.
In this embodiment, the communication protocol corresponding to the video may include: RTP (Real-Time Transport Protocol), RTSP (Real Time Streaming Protocol), RTMP (Real Time Messaging Protocol), and the like. It is understood that the embodiment of the present application does not impose a limitation on the specific communication protocol corresponding to the video. The communication protocol can be adopted between the terminal and the acquisition device.
In an optional embodiment of the present application, the resolution corresponding to the video may specifically include: a first resolution, or a second resolution; wherein the second resolution may be less than the first resolution.
Resolution may refer to the number of pixel points contained in a unit of inch. Generally, the higher the resolution of the video, the more pixels the video contains, the sharper the picture of the video, and the more storage space it occupies.
The embodiment of the application can provide videos with different resolutions, so that a person or a user in the art can select the video with the required resolution to display according to the actual application requirement.
In an optional embodiment of the present application, the step 201 of acquiring a video corresponding to a conference environment may specifically include: collecting a video corresponding to a first resolution ratio aiming at a conference environment; the method may further include: and carrying out fuzzy processing on the video corresponding to the first resolution so as to obtain the video corresponding to the second resolution. The blurring processing can reduce the resolution of the video, and further can reduce the storage space occupied by the video. It should be noted that the video corresponding to the second resolution after the blurring processing still can express the video attribute, so that the video attribute of the interface can be increased, and the probability of the misoperation can be reduced.
The blurring process may include: mean fuzzy processing, gaussian fuzzy processing, and the like. The principle of the blurring process may be: calculating the accumulated sum of certain characteristic values of related pixels and corresponding weights in a certain field around a pixel, and then obtaining a result value; the method can make part of pixel points of the video image lose details and realize fuzzy processing. For example, the weight of the pixel of the mean value blurring process is the same, and the weight of the gaussian blurring process and the distance of the pixel from the center point are in gaussian distribution.
The operator interface receiving the video conference invitation may be used to display the invitation to the video conference, and the corresponding prompt text may include "video conference invitation from user a" and so on. According to the embodiment of the application, the video corresponding to the conference environment is displayed in the operation interface for receiving the video conference invitation, and the video attribute of the invitation interface can be increased.
In this embodiment of the application, optionally, the operation interface for receiving the video conference invitation may include: the first display layer and the second display layer positioned above the first display layer can display videos corresponding to the conference environment on the first display layer, and can display prompt texts, answering controls and rejection controls on the second display layer. The embodiment of the application can display the video as the background corresponding to the operation interface, and the video as the background can not influence the content of the foreground, for example, the operation entrance of the foreground can not be influenced.
Referring to fig. 3, a schematic diagram of an operation interface for receiving a video conference invitation according to an embodiment of the present application is shown, where the operation interface may specifically include: a video 301 corresponding to the conference environment in the background layer, and prompt text 302, a listening control 303, and a rejection control 304 in the foreground layer. Since the video 301 can add the video attribute of the operation interface, the user can be prompted to use the function corresponding to the video in the operation interface, that is, the video conference can be developed. The user can also be helped to judge which one of the answering control 303 and the rejecting control 304 in the interface is operated, so that the probability of misoperation can be reduced.
The operation interface for initiating the video conference can comprise: an initiating entry for a videoconference; the initiating portal for the videoconference can be used to initiate the videoconference. The operation interface for initiating the video conference can be a home interface of a video conference platform, or a desktop interface of a terminal, and the like, and the types of the video platform can include: APP (Application), website, or the like. It is understood that the embodiment of the present application does not impose a limitation on the specific interface where the originating portal is located.
It is understood that the operation interface for initiating the video conference may include other entries besides the entry for initiating the video conference, such as an entry for initiating a conference call, an entry for an address book, an entry for a dial, and the like.
Referring to fig. 4, a schematic diagram of an operation interface for initiating a video conference according to an embodiment of the present application is shown, where the operation interface may specifically include: an initiating entry 401 of a video conference, an initiating entry 402 of a teleconference, an address book 403, a dial 404, and the like. The video corresponding to the conference environment is displayed in the initiating entry 401 of the video conference, so that the video attribute of the initiating entry 401 of the video conference can be increased, and a user can be prompted to use the function corresponding to the initiating entry 401 of the video conference to expand the video conference. The user can also be helped to judge whether to operate on the initiating entrance 401 of the video conference, so that the probability of misoperation can be reduced.
Optionally, the operation interface for initiating the video conference may include: an initiating portal for a videoconference, the method may further comprise: and initiating a video conference in response to the triggering operation aiming at the initiating entrance.
Optionally, the process of initiating the video conference may include: determining a communication opposite end; and sending video conference invitation to the correspondent node. It is to be understood that the embodiment of the present application does not impose a limitation on the specific process of initiating the video conference.
It can be understood that the interface where the initiating entry of the video conference is located and the operation interface for receiving the video conference invitation are only optional embodiments of the operation interface in the embodiment of the present application, and actually, the operation interface in the embodiment of the present application may be any operation interface related to the video conference, and the embodiment of the present application does not limit a specific interface.
According to an embodiment, the acquisition device corresponding to the conference environment can be always in an on state, and then the video corresponding to the conference environment can be continuously acquired through the acquisition device and displayed.
According to another embodiment, the acquisition device corresponding to the conference environment may be in a closed state, and in this case, the video corresponding to the conference environment may not be acquired and displayed. Under the condition that the starting condition is met, the acquisition device is started to be in a starting state, and videos corresponding to the conference environment are acquired and displayed; the starting condition may be related to a video conference requirement of the user, for example, when the user has the video conference requirement, a video corresponding to the conference environment is displayed to the user, so that the intelligence of interface processing may be improved. For example, the terminal is a user device, and when the user is not present in the conference environment, the video corresponding to the conference environment may not be displayed in the interface of the user terminal; and when the user is in the conference environment, the video corresponding to the conference environment can be displayed in the interface of the user terminal.
Alternatively, the turn-on condition may include: people exist in the conference environment, and under the condition, the user can be considered to have the video conference requirement, so that the acquisition device corresponding to the conference environment can be started. Optionally, whether the opening condition is met or not may be determined according to the user state information of one dimension or multiple dimensions, which will be described in detail in method embodiment two.
In an optional embodiment of the present application, the conference environment may correspond to N acquisition devices, where N may be a natural number, N is equal to 1, or N is greater than 1, that is, 1 or more acquisition devices may be disposed in the conference environment; the video can be acquired by a target acquisition device, and the target acquisition device can correspond to all or part of the N acquisition devices.
In practical application, a part can be determined from the N acquisition devices according to the scheduling characteristics and used as a target acquisition device. The scheduling features may include: the performance characteristics of the acquisition device, the position relationship characteristics between the acquisition device and the terminal, the position relationship characteristics between the acquisition device and the entrance of the conference environment, and the like.
The performance characteristics may characterize the acquisition performance of the acquisition device, such as resolution, illumination, and the like. The positional relationship features may include: distance. Optionally, a collection device closer to an entrance of the conference environment may be preferentially selected, or a collection device with better performance characteristics may be preferentially selected, or a collection device closer to the terminal may be preferentially selected.
Of course, the above-mentioned positional relationship characteristics may include, in addition to the distance: relative direction, and the like, and the specific positional relationship characteristics are not limited in the embodiments of the present application.
In an optional embodiment of the present application, the number of the target acquisition devices may be multiple, and the video corresponding to the conference environment is a multi-picture video obtained by synthesizing video data acquired by multiple target acquisition devices.
The embodiment of the application can synthesize the video data acquired by the plurality of target acquisition devices to obtain a multi-picture video, so that the conditions of the conference environment acquired by the plurality of visual angles corresponding to the plurality of target acquisition devices can be displayed in real time. Alternatively, the number of pictures of the multi-picture video may match the number of target capture devices.
In step 202, a video display technology may be used to display a video corresponding to the conference environment in an operation interface. Optionally, a video corresponding to the conference environment may be displayed in the operation interface by using HTML5 (HyperText Markup Language version 5) Stream technology. H5Stream provides an out-of-box Streaming server function, can configure the source of video, and can start, pause, or stop video in a browser. The H5Stream technique may include: media Source Extensions (MSE) technology, or Webrtc (Web Real-Time Communication) technology, etc.
In an optional embodiment of the present Application, an API (Application Programming Interface) provided by Webrtc may be utilized to configure a source of a video, and display a video corresponding to the conference environment in an operation Interface of a browser. It can be understood that, the embodiment of the present application does not impose a limitation on the specific process of displaying the video corresponding to the conference environment in the operation interface.
In summary, in the data processing method according to the embodiment of the present application, the video corresponding to the conference environment is displayed in the operation interface, and the video attribute of the operation interface can be added to the video in the operation interface, so that more bases can be provided for understanding the interface function, and specifically, a user can be prompted to use the function corresponding to the video in the operation interface to expand the video conference.
In addition, as the video attribute of the interface is added to the video in the operation interface, more bases can be provided for the operation corresponding to the operation interface, and the user can be prompted to operate the operation interface, so that the probability of misoperation can be reduced.
Method embodiment two
Referring to fig. 5, a flowchart illustrating steps of a second embodiment of the data processing method in the present application is shown, which may specifically include the following steps:
step 501, starting a target acquisition device corresponding to a conference environment under the condition that the user state information meets the starting condition;
502, acquiring a video corresponding to the conference environment through the target acquisition device;
step 503, displaying a video corresponding to the conference environment in an operation interface;
the operation interface may specifically include:
receiving an operation interface of the video conference invitation; or
And initiating an operation interface of the video conference.
In the embodiment of the application, the target acquisition device corresponding to the conference environment can be in a closed state by default, and the target acquisition device can be opened to be in an opened state under the condition that the user state meets the opening condition, so that the video corresponding to the conference environment can be displayed in the interface. Because the user state information meets the starting condition, the fact that the user has the video conference requirement can be represented, therefore, in the embodiment of the application, under the condition that the user has the video conference requirement, the source corresponding to the video is started, the video corresponding to the conference environment is displayed in the interface, and the intelligence of interface processing can be improved.
In this embodiment of the application, optionally, the starting condition may include: there is a person in the conference environment. Under the condition that people exist in the conference environment, the fact that the user has the video conference requirement can be determined, and therefore the source corresponding to the video can be started.
The user state information can be the state presented by the user, and the starting condition can be used for representing the condition corresponding to the starting of the target acquisition device; according to the method and the device, the target acquisition device can be started by the user through the presented user state information, so that the intelligence of interface processing can be improved.
The user status information may include at least one of the following information:
user face information; the user face information may include: eye, lip, or nose information.
The user limb information can refer to specific actions and body positions of the user in the process of applying limbs;
location information of the terminal; in the case where the terminal is a user device, the location information of the terminal may be used as a basis for whether the user is located in the conference environment.
Position relation information between the user and the terminal; and
and sound information.
The positional relationship information may include: distance information, or relative direction information between the user and the terminal. The relative direction information may include: a first direction or a second direction.
Wherein, the first direction may be facing or approximately facing; facing may refer to the user's face facing a preset face of the device. The second direction may be back, or approximately back; back-facing may refer to the user's back facing a preset face of the device. The preset surface can be determined by those skilled in the art according to the actual application requirements, for example, the preset surface can be set by a user, or the preset surface can include: the face towards which the user's face is oriented during execution of the preset function.
In an alternative embodiment of the present application, user status information may be determined from data collected by the sensors.
The sensor may include: an image sensor for capturing an image to locate a user's face from the image data, thereby enabling determination of user status information, such as user facial information.
The sensor may include: a distance sensor, etc. The distance sensor may detect a distance between two objects based on light waves, sound waves, or electromagnetic waves.
For example, the data collected by the sensors may include: image data, which may be captured by an image sensor such as a camera, video camera, etc. Alternatively, the number of image sensors may be 1, or may be greater than 1. Alternatively, the image sensor may be disposed inside or outside the terminal.
Alternatively, at least one of user face information, user limb information, and positional relationship information between the user and the terminal may be determined from the image data. For example, the front of the user (e.g., face or chest) is included in the image data, and the relative direction information between the user and the terminal is the first direction. As another example, the image data includes the back of the user (e.g., the back or the back brain), and the relative direction information between the user and the terminal is the second direction.
Optionally, the user status information may specifically include: a lip feature sequence, which may include: an associated plurality of lip features.
In an optional embodiment of the present application, the process of determining the user status information may specifically include: determining a plurality of frame images; the multi-frame images can be images acquired from front to back according to the time sequence; determining a lip feature sequence according to the lip features corresponding to the multi-frame images; the order of the lip features in the lip feature sequence may correspond to the order of the multi-frame images.
The limb information may include: gestures corresponding to the upper limbs or information corresponding to the lower limbs.
The gesture may refer to a posture of an upper limb, and specifically may refer to a specific action and a specific body position of the user when the user operates an arm. According to the gesture recognition device, the gesture of the user can be collected, and the collected gesture is recognized. The gesture recognition is used for determining which gesture belongs, such as determining whether the gesture is a gesture corresponding to the opening condition.
In this embodiment, the orientation of the gesture may include: at least one of a direction, an angle, and an arc. In practical applications, the hand in the preset state may move along any direction, or the hand in the preset state may rotate around a certain point in a plane, or the hand in the preset state may turn around a certain axis, so as to obtain various gestures. Wherein the preset state may include: a fist-making state, a palm unfolding state, a palm half unfolding state and the like. It can be understood that, according to the actual application requirements, a person skilled in the art or a user may set a gesture corresponding to the opening condition, and use the required information to characterize the direction of the gesture, and the embodiment of the present application does not limit the specific gesture and the specific direction representing method of the gesture.
In an application example of the present application, the gesture recognition apparatus may include: the device comprises two infrared cameras, an image sensor, an infrared light emitting diode driving unit, an operation unit and a gesture recognition unit. The image sensor is electrically connected with the two infrared cameras, the infrared light-emitting diode driving unit is electrically connected with the infrared light-emitting diodes to drive the infrared light-emitting diodes to emit light, and the operation unit is electrically connected with the image sensor, the infrared light-emitting diode driving unit and the gesture recognition unit; the image sensor collects gesture image data through the two infrared cameras and sends the gesture image data to the operation unit for processing, and the operation unit obtains the gesture image data, carries out three-dimensional coordinate positioning and then sends the gesture image data to the gesture recognition unit for pattern matching so as to realize gesture recognition.
In practical applications, the gesture recognition apparatus may be independent from a terminal for performing the method of the embodiment of the present application, or the gesture recognition apparatus may be integrated in the terminal for performing the method of the embodiment of the present application. It is understood that a person skilled in the art may implement gesture recognition according to any gesture recognition device as required, and the embodiments of the present application are not limited to a specific gesture recognition device and a specific relationship between the gesture recognition device and a device for performing the method of the embodiments of the present application.
The sound information may refer to collected sound information. For terminals disposed within a conference environment, the sound information may include: audio information within a conference environment.
The sound parameters corresponding to the sound information may include at least one of the following parameters: sound properties, presence of sound and sound loudness.
The sound attributes may include: speech (i.e., human voice), a footstep sound, a percussive sound (e.g., closing a door), or a tap sound (e.g., tapping a table, etc.).
The occurrence of sound may include: duration of sound, etc.
The loudness of sound is also called volume. The loudness of sound may refer to the intensity of sound perceived by the human ear, which is a subjective perception of the magnitude of sound by humans. The loudness of a sound is determined by the amplitude at which the sound is received.
The above-mentioned starting conditions can be determined by those skilled in the art according to the requirements of practical application.
Optionally, the user state information meets the starting condition, which specifically includes:
detecting user face information, and under the condition that the terminal is arranged in a conference environment, if the user face information is detected, indicating that a person is in the conference environment; and/or
Detecting the limb information of the user, and under the condition that the terminal is arranged in the conference environment, if the limb information of the user is detected, indicating that a person is in the conference environment; and/or
Detecting sound information, and under the condition that the terminal is arranged in the conference environment, if the sound information with certain volume is detected, indicating that a person is in the conference environment; and/or
The user face information is preset face information; the preset face information can be triggered by a user to realize the intelligent starting of the target acquisition device; and/or
The user limb information is preset limb information; the preset limb information can be triggered by a user to realize intelligent starting of the target acquisition device; and/or
The position information of the terminal is matched with the position information of the conference environment, and under the condition that the terminal is user equipment, the situation that the user is located in the conference environment can be described; and/or
The distance information between the user and the terminal does not exceed a distance threshold; and/or
The distance information between the user and the terminal represents that the user approaches the terminal; and/or
The relative direction information between the user and the terminal is a first direction; and/or
And the sound parameters corresponding to the sound information accord with the sound starting conditions.
The distance information between the user and the terminal can represent whether the user has a video conference requirement. Generally, the information of the distance between the user and the terminal indicates that the user approaches the terminal gradually, which indicates that the user has an intention to use the terminal, so that the user can be considered to have a video conference requirement. Otherwise, the distance information between the user and the terminal represents that the user gradually gets away from the terminal, which indicates that the user does not have the video conference requirement.
The relative direction information may characterize whether the user has a video conference need. For example, if the front of the user faces a preset face of the terminal, the user may be considered to have a video conference need. Otherwise, the back of the user faces the preset surface of the terminal, which indicates that the user does not have the video conference requirement.
The sound turn-on condition can be determined by those skilled in the art according to the actual application requirements.
Optionally, the sound parameter corresponding to the sound information meets the sound starting condition, and may specifically include at least one of the following conditions:
detecting the voice of a preset volume; and
and detecting the footstep sound with preset time.
In an application example 1 of the present application, a human face or a body trunk structure in a conference environment is captured, and then it can be considered that the opening condition is met.
In an application example 2 of the present application, a microphone in a conference environment receives a sound, and determines whether an open condition is met according to multiple dimensions of sound attributes (which may be human voice, footstep sound, impact sound, etc.), presence, and loudness of the sound.
In an application example 3 of the present application, after a face, a limb, and a voice are simultaneously recognized, it is determined whether an opening condition is satisfied.
In summary, the data processing method of the embodiment of the application judges whether the conference environment is manned or not through the user state information of one dimension or multiple dimensions, and if yes, the video acquired by the acquisition device in the conference environment can be acquired in real time and displayed in real time in the interface, so that the video attribute of the interface can be increased, and the prompt effect on the operation of the user can be realized.
It should be noted that, in the case that the target collecting device is in the on state and the off state, the display content of the same interface may be different. Specifically, under the condition that the target collection device is in an open state, the display content of the interface may include: video corresponding to a conference environment; and under the condition that the target acquisition device is in the closed state, the display content of the interface may not include: video corresponding to a conference environment.
The video conference system and the video conference method can respond to the operation of a user on the interface and expand the video conference. For example, a video conference may be joined in response to a user operating an answer control in an invitation interface. In another example, the video conference may be initiated in response to a user operation of an initiating portal in an interface in which the initiating portal of the video conference is located.
After the video conference is developed, the terminal can be used as a conference device, and can acquire data such as voice, video and the like of a user in a conference environment according to the flow shown in fig. 1, and the data is sent to the control device through a network in a format of being preprocessed, compressed and encoded, and packaged into a data packet; and meanwhile, a data packet at a far end is received from the network, and effective data obtained after the data packet is unpacked is decoded and then played to a user.
Referring to fig. 6, a schematic structural diagram of a video conference system according to an embodiment of the present application is shown, where the system is implemented based on a Webrtc technology, and specifically may include: at least two conference terminals 601, a signaling server 602 and a pass through server 603.
A web client runs on the conference terminal 601, and calls a collection device through the Webrtc characteristic of the HTML5 to obtain media data such as voice, video and the like of a user in a conference environment; the Webrtc characteristic can also be utilized to process the media data to obtain a media data packet, and the media data packet is transmitted to other conference terminals 601 based on the point-to-point connection.
Point-to-point connections may be established between different conference terminals 601, and the types of point-to-point connections may include: webrtc, Web Socket, TCP (Transmission Control Protocol), or the like.
The signaling server 602 is a role for managing and assisting the conference terminals 601 to establish a point-to-point connection to the center, and the signaling server 602 transfers information of each other between different conference terminals 601 so that different conference terminals 601 establish a point-to-point connection through the information and transfer media data through the point-to-point connection. The roles of this role may include:
a connection control message for controlling initiation or termination of a communication;
messages for mutual notification when errors occur;
media stream metadata on the respective side, such as decoder, configuration of decoder, bandwidth, media type, etc.;
key data used for establishing safe connection between every two devices;
data on the network that can be seen from the outside world, such as a wide area network IP (Internet Protocol Address), a port, and the like.
The penetration server 603 may function as an intranet penetration.
It is understood that the video conference system implemented based on Webrtc technology in fig. 6 is only an alternative embodiment, and actually, those skilled in the art may implement video conference by using other technologies according to the actual application requirements, and the other technologies may include: VOIP (Voice over IP) technology, and the like, and the embodiment of the present application does not limit the specific technology for implementing the video conference.
Method example III
Referring to fig. 7, a flowchart illustrating steps of a third embodiment of the data processing method according to the present application is shown, where the method may specifically include the following steps:
701, collecting a video corresponding to a conference environment;
step 702, displaying the video as a background of an operation interface;
the operation interface may specifically include:
receiving an operation interface of the video conference invitation; or
And initiating an operation interface of the video conference.
The video is displayed as the background corresponding to the operation interface, and the video serving as the background can not influence the content of the foreground, for example, the operation entrance of the foreground can not be influenced.
In this embodiment of the present application, optionally, the operation interface may include: the first display layer and the second display layer positioned above the first display layer can display the video corresponding to the conference environment on the first display layer, and can display the original content, such as prompt text, operation entry and the like, on the second display layer.
Optionally, the operation interface for initiating the video conference may include: an initiating portal for a videoconference, the method may further comprise: and initiating a video conference in response to the triggering operation aiming at the initiating entrance.
Optionally, the process of initiating the video conference may include: determining a communication opposite end; and sending video conference invitation to the correspondent node. It is to be understood that the embodiment of the present application does not impose a limitation on the specific process of initiating the video conference.
It should be noted that, for simplicity of description, the method embodiments are described as a series of acts or combination of acts, but those skilled in the art will recognize that the embodiments are not limited by the order of acts described, as some steps may occur in other orders or concurrently depending on the embodiments. Further, those of skill in the art will recognize that the embodiments described in this specification are presently preferred embodiments and that no particular act is required to implement the embodiments of the disclosure.
The embodiment of the application also provides a data processing device.
Referring to fig. 8, a block diagram of a data processing apparatus according to an embodiment of the present application is shown, which may specifically include the following modules:
the video acquisition module 801 is used for acquiring videos corresponding to the conference environment; and
a video display module 802, configured to display a video corresponding to the conference environment in an operation interface;
the operation interface may specifically include:
receiving an operation interface of the video conference invitation; or
And initiating an operation interface of the video conference.
Optionally, the operation interface for initiating the video conference includes: an initiating portal for a videoconference, the apparatus may further comprise:
and the initiating module is used for responding to the triggering operation aiming at the initiating entrance and initiating the video conference.
Optionally, the resolution corresponding to the video may specifically include:
a first resolution, or a second resolution; wherein the second resolution is less than the first resolution.
Optionally, the video acquisition module 801 is specifically configured to acquire a video corresponding to the first resolution for the conference environment;
the apparatus may further include:
and the fuzzy processing module is used for carrying out fuzzy processing on the video corresponding to the first resolution so as to obtain the video corresponding to the second resolution.
Optionally, the video capture module 801 may specifically include:
the starting module is used for starting the target acquisition device corresponding to the conference environment under the condition that the user state information accords with the starting condition;
and the acquisition module is used for acquiring the video corresponding to the conference environment through the target acquisition device.
Optionally, the user status information may include at least one of the following information:
user face information;
user limb information;
location information of the terminal;
information of a position relationship between a user and the terminal; and
and (4) sound information.
Optionally, the user state information meets the start condition, and may include:
detecting user face information; and/or
Detecting user limb information; and/or
Detecting sound information; and/or
The user face information is preset face information; and/or
The user limb information is preset limb information; and/or
The position information of the terminal is matched with the position information of the conference environment; and/or
The distance information between the user and the terminal does not exceed a distance threshold; and/or
The distance information between the user and the terminal represents that the user approaches the terminal; and/or
The relative direction information between the user and the terminal is a first direction; and/or
And the sound parameters corresponding to the sound information accord with the sound starting conditions.
Optionally, the sound parameters may include at least one of the following parameters:
sound properties, presence of sound and sound loudness.
Optionally, the conference environment corresponds to N acquisition devices, where N is a natural number;
the video can be acquired by a target acquisition device, and the target acquisition device can correspond to all or part of the N acquisition devices.
Optionally, the video corresponding to the conference environment may be a multi-picture video obtained by synthesizing video data acquired by a plurality of target acquisition devices.
Referring to fig. 9, a block diagram of a data processing apparatus according to an embodiment of the present application is shown, which may specifically include the following modules:
the video acquisition module 901 is used for acquiring videos corresponding to the conference environment; and
a video display module 902, configured to display the video as a background of the operation interface;
the operation interface may specifically include:
receiving an operation interface of the video conference invitation; or
And initiating an operation interface of the video conference.
Optionally, the operation interface for initiating the video conference includes: an initiating portal for a videoconference, the apparatus may further comprise:
and the initiating module is used for responding to the triggering operation aiming at the initiating entrance and initiating the video conference.
Optionally, the resolution corresponding to the video may specifically include:
a first resolution, or a second resolution; wherein the second resolution is less than the first resolution.
Optionally, the video acquisition module 901 is specifically configured to acquire a video corresponding to the first resolution for the conference environment;
the apparatus may further include:
and the fuzzy processing module is used for carrying out fuzzy processing on the video corresponding to the first resolution so as to obtain the video corresponding to the second resolution.
Optionally, the video capturing module 901 may specifically include:
the starting module is used for starting the target acquisition device corresponding to the conference environment under the condition that the user state information accords with the starting condition;
and the acquisition module is used for acquiring the video corresponding to the conference environment through the target acquisition device.
Optionally, the user status information may include at least one of the following information:
user face information;
user limb information;
location information of the terminal;
information of a positional relationship between a user and the terminal; and
and sound information.
Optionally, the user state information meets the start condition, and may include:
detecting user face information; and/or
Detecting user limb information; and/or
Detecting sound information; and/or
The user face information is preset face information; and/or
The user limb information is preset limb information; and/or
The position information of the terminal is matched with the position information of the conference environment; and/or
The distance information between the user and the terminal does not exceed a distance threshold; and/or
The distance information between the user and the terminal represents that the user approaches the terminal; and/or
The relative direction information between the user and the terminal is a first direction; and/or
And the sound parameters corresponding to the sound information accord with the sound starting conditions.
Optionally, the sound parameters may include at least one of the following parameters:
sound properties, presence of sound and sound loudness.
Optionally, the conference environment corresponds to N acquisition devices, where N is a natural number;
the video can be acquired by a target acquisition device, and the target acquisition device can correspond to all or part of the N acquisition devices.
Optionally, the video corresponding to the conference environment may be a multi-picture video obtained by synthesizing video data acquired by a plurality of target acquisition devices.
The embodiments in the present specification are described in a progressive manner, each embodiment focuses on differences from other embodiments, and the same and similar parts among the embodiments are referred to each other.
With regard to the apparatus in the above-described embodiment, the specific manner in which each module performs the operation has been described in detail in the embodiment related to the method, and will not be elaborated here.
Embodiments of the application can be implemented as a system or apparatus employing any suitable hardware and/or software for the desired configuration. Fig. 10 schematically illustrates an example device 1300 that can be used to implement various embodiments described herein.
For one embodiment, fig. 10 illustrates an exemplary apparatus 1300, which apparatus 1300 may comprise: one or more processors 1302, a system control module (chipset) 1304 coupled to at least one of the processors 1302, system memory 1306 coupled to the system control module 1304, non-volatile memory (NVM)/storage 1308 coupled to the system control module 1304, one or more input/output devices 1310 coupled to the system control module 1304, and a network interface 1312 coupled to the system control module 1306. The system memory 1306 may include: instruction 1362, the instruction 1362 executable by the one or more processors 1302.
Processor 1302 may include one or more single-core or multi-core processors, and processor 1302 may include any combination of general-purpose processors or special-purpose processors (e.g., graphics processors, application processors, baseband processors, etc.). In some embodiments, the device 1300 can be a server, a target device, a wireless device, etc. as described in embodiments herein.
In some embodiments, device 1300 may include one or more machine-readable media (e.g., system memory 1306 or NVM/storage 1308) having instructions and one or more processors 1302, which may be configured to execute the instructions to implement the modules included by the aforementioned means, thereby performing the actions described in embodiments of the present application, in conjunction with the one or more machine-readable media.
System control module 1304 for one embodiment may include any suitable interface controller to provide any suitable interface to at least one of processors 1302 and/or any suitable device or component in communication with system control module 1304.
System control module 1304 for one embodiment may include one or more memory controllers to provide an interface to system memory 1306. The memory controller may be a hardware module, a software module, and/or a firmware module.
System memory 1306 for one embodiment may be used to load and store data and/or instructions 1362. For one embodiment, system memory 1306 may include any suitable volatile memory, such as suitable DRAM (dynamic random access memory). In some embodiments, system memory 1306 may include: double data rate type four synchronous dynamic random access memory (DDR4 SDRAM).
System control module 1304 for one embodiment may include one or more input/output controllers to provide an interface to NVM/storage 1308 and input/output device(s) 1310.
NVM/storage 1308 for one embodiment may be used to store data and/or instructions 1382. NVM/storage 1308 may include any suitable non-volatile memory (e.g., flash memory, etc.) and/or may include any suitable non-volatile storage device(s), e.g., one or more Hard Disk Drives (HDDs), one or more Compact Disc (CD) drives, and/or one or more Digital Versatile Disc (DVD) drives, etc.
The NVM/storage 1308 may include storage resources that are physically part of the device on which the apparatus 1300 is installed or may be accessible by the device and not necessarily part of the device. For example, the NVM/storage 1308 may be accessible over a network via the network interface 1312 and/or through the input/output devices 1310.
Input/output device(s) 1310 for one embodiment may provide an interface for apparatus 1300 to communicate with any other suitable device, and input/output devices 1310 may include communication components, audio components, sensor components, and so forth.
Network interface 1312 of one embodiment may provide an interface for device 1300 to communicate with one or more networks and/or with any other suitable apparatus, and device 1300 may communicate wirelessly with one or more components of a wireless network according to any of one or more wireless network standards and/or protocols, such as to access a communication standard-based wireless network, such as WiFi, 2G, or 3G, or a combination thereof.
For one embodiment, at least one of the processors 1302 may be packaged together with logic for one or more controllers (e.g., memory controllers) of the system control module 1304. For one embodiment, at least one of processors 1302 may be packaged together with logic for one or more controllers of system control module 1304 to form a System In Package (SiP). For one embodiment, at least one of the processors 1302 may be integrated on the same novelty as the logic of one or more controllers of the system control module 1304. For one embodiment, at least one of processors 1302 may be integrated on the same chip with logic for one or more controllers of system control module 1304 to form a system on a chip (SoC).
In various embodiments, apparatus 1300 may include, but is not limited to: a computing device such as a desktop computing device or a mobile computing device (e.g., a laptop computing device, a handheld computing device, a tablet, a netbook, etc.). In various embodiments, device 1300 may have more or fewer components and/or different architectures. For example, in some embodiments, device 1300 may include one or more cameras, a keyboard, a Liquid Crystal Display (LCD) screen (including a touch screen display), a non-volatile memory port, multiple antennas, a graphics chip, an Application Specific Integrated Circuit (ASIC), and speakers.
Wherein, if the display includes a touch panel, the display screen may be implemented as a touch screen display to receive an input signal from a user. The touch panel includes one or more touch sensors to sense touch, slide, and gestures on the touch panel. The touch sensor may not only sense the boundary of a touch or slide action, but also detect the duration and pressure associated with the touch or slide operation.
The present application also provides a non-transitory readable storage medium, where one or more modules (programs) are stored in the storage medium, and when the one or more modules are applied to an apparatus, the apparatus may be caused to execute instructions (instructions) of methods in the present application.
Provided in one example is an apparatus comprising: one or more processors; and, instructions in one or more machine-readable media stored thereon, which when executed by the one or more processors, cause the apparatus to perform a method as in embodiments of the present application, which may include: the method shown in fig. 2 or fig. 3 or fig. 4 or fig. 5 or fig. 6 or fig. 7.
One or more machine-readable media are also provided in one example, having instructions stored thereon, which when executed by one or more processors, cause an apparatus to perform a method as in embodiments of the application, which may include: the method of fig. 2 or fig. 3 or fig. 4 or fig. 5 or fig. 6 or fig. 7.
The specific manner in which each module performs operations of the apparatus in the above embodiments has been described in detail in the embodiments related to the method, and will not be described in detail here, and reference may be made to part of the description of the method embodiments for relevant points.
The embodiments in the present specification are described in a progressive manner, each embodiment focuses on differences from other embodiments, and the same and similar parts among the embodiments are referred to each other.
Embodiments of the present application are described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the application. It will be understood that each flow and/or block of the flow diagrams and/or block diagrams, and combinations of flows and/or blocks in the flow diagrams and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable conferencing terminal to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable conferencing terminal, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable conferencing terminal to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable conferencing terminal to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
While preferred embodiments of the present application have been described, additional variations and modifications of these embodiments may occur to those skilled in the art once they learn of the basic inventive concepts. Therefore, it is intended that the appended claims be interpreted as including the preferred embodiment and all such alterations and modifications as fall within the true scope of the embodiments of the application.
Finally, it should also be noted that, herein, relational terms such as first and second, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. Also, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other identical elements in a process, method, article, or apparatus that comprises the element.
The foregoing detailed description has provided a data processing method, a data processing apparatus, a device, and a machine-readable medium, and the principles and embodiments of the present application have been described herein using specific examples, which are provided only to help understand the method and the core ideas of the present application; meanwhile, for a person skilled in the art, according to the idea of the present application, there may be variations in the specific embodiments and the application scope, and in summary, the content of the present specification should not be construed as a limitation to the present application.

Claims (18)

1. A data processing method, comprising:
collecting a video corresponding to a conference environment;
displaying a video corresponding to the conference environment in an operation interface;
wherein, the operation interface includes:
receiving an operation interface of the video conference invitation; before receiving a video conference invitation, displaying a video corresponding to a conference environment of the terminal in an operation interface; or
Initiating an operation interface of the video conference; the operation interface for initiating the video conference comprises the following steps: an initiating entry for a videoconference; and before the video conference is initiated through the initiating entrance, displaying a video corresponding to the conference environment of the terminal in the initiating entrance.
2. The method of claim 1, further comprising:
and initiating a video conference in response to the triggering operation aiming at the initiating entrance.
3. The method of claim 1, wherein the corresponding resolution of the video comprises:
a first resolution, or a second resolution; wherein the second resolution is less than the first resolution.
4. The method of claim 3, wherein capturing video corresponding to a meeting environment comprises:
acquiring a video corresponding to a first resolution ratio aiming at a conference environment;
the method further comprises the following steps: and carrying out fuzzy processing on the video corresponding to the first resolution so as to obtain the video corresponding to the second resolution.
5. The method of any one of claims 1 to 4, wherein capturing video corresponding to a meeting environment comprises:
starting a target acquisition device corresponding to the conference environment under the condition that the user state information meets the starting condition;
and acquiring the video corresponding to the conference environment through the target acquisition device.
6. The method of claim 5, wherein the user status information comprises at least one of the following information:
user face information;
user limb information;
location information of the terminal;
information of a positional relationship between a user and the terminal; and
and sound information.
7. The method of claim 5, wherein the user status information complies with a turn-on condition, comprising:
detecting user face information; and/or
Detecting user limb information; and/or
Detecting sound information; and/or
The user face information is preset face information; and/or
The user limb information is preset limb information; and/or
The position information of the terminal is matched with the position information of the conference environment; and/or
The distance information between the user and the terminal does not exceed a distance threshold; and/or
The distance information between the user and the terminal represents that the user approaches the terminal; and/or
The relative direction information between the user and the terminal is a first direction; and/or
And the sound parameters corresponding to the sound information accord with the sound starting conditions.
8. The method of claim 7, wherein the sound parameters comprise at least one of:
sound properties, presence of sound and sound loudness.
9. The method according to any one of claims 1 to 4, wherein the conference environment corresponds to N acquisition devices, N being a natural number;
the videos are acquired by target acquisition devices, and the target acquisition devices correspond to all or part of the N acquisition devices.
10. The method according to any one of claims 1 to 4, wherein the video corresponding to the conference environment is a multi-picture video obtained by synthesizing video data acquired by a plurality of target acquisition devices.
11. A data processing method, comprising:
collecting a video corresponding to a conference environment;
displaying the video as a background of an operation interface;
wherein, the operation interface includes:
receiving an operation interface of the video conference invitation; before receiving a video conference invitation, displaying a video corresponding to a conference environment of the terminal in an operation interface; or alternatively
Initiating an operation interface of the video conference; the operation interface for initiating the video conference comprises the following steps: an initiating entry for a videoconference; and before the video conference is initiated through the initiating entrance, displaying a video corresponding to the conference environment of the terminal in the initiating entrance.
12. The method of claim 11, wherein the operation interface for initiating the video conference comprises: an initiating portal for a videoconference, the method further comprising:
and initiating a video conference in response to the triggering operation aiming at the initiating entrance.
13. A data processing apparatus, comprising:
the video acquisition module is used for determining a video corresponding to the conference environment; and
the video display module is used for displaying the video corresponding to the conference environment in an operation interface;
wherein, the operation interface includes:
receiving an operation interface of the video conference invitation; before receiving a video conference invitation, displaying a video corresponding to a conference environment of the terminal in an operation interface; or
Initiating an operation interface of the video conference; the operation interface for initiating the video conference comprises the following steps: an initiating entry for a videoconference; and before the video conference is initiated through the initiating entrance, displaying a video corresponding to the conference environment of the terminal in the initiating entrance.
14. A data processing apparatus, comprising:
the video acquisition module is used for determining a video corresponding to the conference environment; and
the video display module is used for displaying the video as the background of the operation interface;
wherein, the operation interface includes:
receiving an operation interface of the video conference invitation; before receiving a video conference invitation, displaying a video corresponding to a conference environment of the terminal in an operation interface; or
Initiating an operation interface of the video conference; the operation interface for initiating the video conference comprises the following steps: an initiating entry for a videoconference; and before the video conference is initiated through the initiating entrance, displaying a video corresponding to the conference environment of the terminal in the initiating entrance.
15. An apparatus, comprising:
one or more processors; and
one or more machine-readable media having instructions stored thereon that, when executed by the one or more processors, cause the apparatus to perform the method recited by one or more of claims 1-10.
16. One or more machine-readable media having instructions stored thereon, which when executed by one or more processors, cause an apparatus to perform the method recited by one or more of claims 1-10.
17. An apparatus, comprising:
one or more processors; and
one or more machine-readable media having instructions stored thereon that, when executed by the one or more processors, cause the apparatus to perform the method of one or more of claims 11-12.
18. One or more machine-readable media having instructions stored thereon, which when executed by one or more processors, cause an apparatus to perform the method recited by one or more of claims 11-12.
CN201910426233.0A 2019-05-21 2019-05-21 Data processing method, device, equipment and machine readable medium Active CN111988555B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910426233.0A CN111988555B (en) 2019-05-21 2019-05-21 Data processing method, device, equipment and machine readable medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910426233.0A CN111988555B (en) 2019-05-21 2019-05-21 Data processing method, device, equipment and machine readable medium

Publications (2)

Publication Number Publication Date
CN111988555A CN111988555A (en) 2020-11-24
CN111988555B true CN111988555B (en) 2022-05-24

Family

ID=73436224

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910426233.0A Active CN111988555B (en) 2019-05-21 2019-05-21 Data processing method, device, equipment and machine readable medium

Country Status (1)

Country Link
CN (1) CN111988555B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112769818A (en) * 2021-01-05 2021-05-07 武汉球之道科技有限公司 Video processing method based on webpage instant messaging and IP communication
CN113347391B (en) * 2021-05-31 2022-12-06 北京字跳网络技术有限公司 Data transmission method, data transmission interruption method and device
CN114422744B (en) * 2021-12-23 2023-06-09 华能招标有限公司 Video picture display method and device for remote comment video conference process
CN117729188B (en) * 2024-02-18 2024-04-12 北京市智慧水务发展研究院 Water affair video acquisition system and method based on WebRTC

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2013067829A1 (en) * 2011-11-08 2013-05-16 华为技术有限公司 Image display adjustment method, device and system
CN108259801A (en) * 2018-01-19 2018-07-06 广州视源电子科技股份有限公司 Audio, video data display methods, device, equipment and storage medium

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
NO332170B1 (en) * 2009-10-14 2012-07-16 Cisco Systems Int Sarl Camera control device and method
US8692862B2 (en) * 2011-02-28 2014-04-08 Cisco Technology, Inc. System and method for selection of video data in a video conference environment
US8917306B2 (en) * 2011-07-29 2014-12-23 Cisco Technology, Inc. Previewing video data in a video communication environment
US8957936B2 (en) * 2012-10-23 2015-02-17 Cisco Technology, Inc. Method to preview caller in a video conference session
WO2016024288A1 (en) * 2014-08-14 2016-02-18 Vats Nitin Realistic viewing and interaction with remote objects or persons during telepresence videoconferencing
US9762729B1 (en) * 2016-05-12 2017-09-12 Google Inc. Caller preview for video calls
CN106850927B (en) * 2016-12-30 2019-05-31 深圳市国华识别科技开发有限公司 Call prompting method and system
CN107222706B (en) * 2017-06-29 2019-12-13 北京奇艺世纪科技有限公司 Video preview method and system
CN108718402B (en) * 2018-08-14 2021-04-13 四川易为智行科技有限公司 Video conference management method and device

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2013067829A1 (en) * 2011-11-08 2013-05-16 华为技术有限公司 Image display adjustment method, device and system
CN108259801A (en) * 2018-01-19 2018-07-06 广州视源电子科技股份有限公司 Audio, video data display methods, device, equipment and storage medium

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
"Vibration-based interface for remote object manipulation in video conference system";T. Ito;《 RO-MAN 2004. 13th IEEE International Workshop on Robot and Human Interactive Communication (IEEE Catalog No.04TH8759)》;20050110;全文 *
张晓坤." 面向虚拟会议系统的图像和视频技术研究".《中国优秀硕士学位论文全文数据库》.2011,全文. *

Also Published As

Publication number Publication date
CN111988555A (en) 2020-11-24

Similar Documents

Publication Publication Date Title
CN111988555B (en) Data processing method, device, equipment and machine readable medium
US8570358B2 (en) Automated wireless three-dimensional (3D) video conferencing via a tunerless television device
US8730295B2 (en) Audio processing for video conferencing
KR101936944B1 (en) Interactive video conferencing
US9143725B2 (en) System and method for providing enhanced graphics in a video environment
US8902244B2 (en) System and method for providing enhanced graphics in a video environment
US20110216153A1 (en) Digital conferencing for mobile devices
CN103222262B (en) For skipping the system and method for Video coding in a network environment
US8923649B2 (en) System and method for calibrating display overscan using a mobile device
US20120056971A1 (en) Virtual Presence Via Mobile
US20120327176A1 (en) Video Call Privacy Control
KR20170091592A (en) Interactive video conferencing
US20140028778A1 (en) Systems and methods for ad-hoc integration of tablets and phones in video communication systems
CN107147927B (en) Live broadcast method and device based on live broadcast wheat connection
CN108293104A (en) Information processing system, wireless terminal and information processing method
US8786631B1 (en) System and method for transferring transparency information in a video environment
KR101863218B1 (en) Video call device and method using webrtc
JP2018515979A (en) Communication processing method and electronic apparatus
CN112203126A (en) Screen projection method, screen projection device and storage medium
JP2015119335A (en) Terminal, system, program and method to thin out frame of photographed moving image in accordance with movement change amount
CN114374880B (en) Joint live broadcast method, joint live broadcast device, electronic equipment and computer readable storage medium
US20210320810A1 (en) Volumetric conversational services using network edge
KR20110138094A (en) Video conference system and method thereof
CN114339140A (en) Interactive monitoring device, video transmission method and device
CN112291507A (en) Video picture adjusting method and device, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
TA01 Transfer of patent application right
TA01 Transfer of patent application right

Effective date of registration: 20201223

Address after: Room 603, 6 / F, Roche Plaza, 788 Cheung Sha Wan Road, Kowloon, China

Applicant after: Zebra smart travel network (Hong Kong) Limited

Address before: A four-storey 847 mailbox in Grand Cayman Capital Building, British Cayman Islands

Applicant before: Alibaba Group Holding Ltd.

GR01 Patent grant
GR01 Patent grant