CN111832455A - Method, device, storage medium and electronic equipment for acquiring content image - Google Patents
Method, device, storage medium and electronic equipment for acquiring content image Download PDFInfo
- Publication number
- CN111832455A CN111832455A CN202010622912.8A CN202010622912A CN111832455A CN 111832455 A CN111832455 A CN 111832455A CN 202010622912 A CN202010622912 A CN 202010622912A CN 111832455 A CN111832455 A CN 111832455A
- Authority
- CN
- China
- Prior art keywords
- image
- content
- monitoring
- images
- target
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 109
- 238000012544 monitoring process Methods 0.000 claims abstract description 163
- 230000008569 process Effects 0.000 claims abstract description 71
- 238000005192 partition Methods 0.000 claims description 24
- 238000012545 processing Methods 0.000 claims description 15
- 238000004590 computer program Methods 0.000 claims description 5
- 230000008859 change Effects 0.000 abstract description 8
- 238000004891 communication Methods 0.000 description 10
- 238000010586 diagram Methods 0.000 description 7
- 238000005516 engineering process Methods 0.000 description 5
- 230000003287 optical effect Effects 0.000 description 4
- 230000005236 sound signal Effects 0.000 description 4
- 230000001133 acceleration Effects 0.000 description 2
- 230000009471 action Effects 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 238000010191 image analysis Methods 0.000 description 2
- 238000003703 image analysis method Methods 0.000 description 2
- 238000003384 imaging method Methods 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 238000007726 management method Methods 0.000 description 2
- 230000006978 adaptation Effects 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 238000013473 artificial intelligence Methods 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000003702 image correction Methods 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000008447 perception Effects 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/10—Terrestrial scenes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/22—Matching criteria, e.g. proximity measures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- Artificial Intelligence (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Bioinformatics & Computational Biology (AREA)
- General Engineering & Computer Science (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
The present disclosure relates to a method, an apparatus, a storage medium, and an electronic device for acquiring a content image, the method including: acquiring a monitoring image containing display content on a target plane at each preset time point after the content display process based on the target plane starts; under the condition that the difference degree between two monitoring images acquired at two adjacent preset time points is larger than a preset difference degree threshold value, generating a content image which is not shielded by a shielding object through a plurality of first monitoring images, wherein the first monitoring images are monitoring images acquired after the target time points with the difference degree larger than the difference degree threshold value are determined; and after the content display process is determined to be terminated, acquiring a plurality of content images corresponding to the content display process. The change of the display content can be sensed in the content display process, and when the display content changes, a complete display content image is extracted from the collected monitoring image, so that the efficiency and the intelligent degree of the display content recording are improved.
Description
Technical Field
The present disclosure relates to the field of artificial intelligence applications, and in particular, to a method, an apparatus, a storage medium, and an electronic device for obtaining a content image.
Background
During a meeting or lecture, a presenter typically presents meeting content or lecture content via presentation content, such as a presentation (e.g., Microsoft Office PowerPoint, PPT) or a blackboard. And the meeting participants or the lecturers usually need to record the content of the presentation or the blackboard writing so as to further study and understand the content of the meeting or the teaching after the meeting or the course. In the related art, the contents of a presentation or a board book are generally recorded by a conventional recording means, for example, a handwriting recording or a means of taking a photograph and a video. The intelligent degree and the recording efficiency of the recording modes are low, and the storage form of the obtained display content is disordered, so that the subsequent reading and understanding of the user are not facilitated.
Disclosure of Invention
To overcome the problems in the related art, the present disclosure provides a method, an apparatus, a storage medium, and an electronic device for acquiring a content image.
According to a first aspect of the embodiments of the present disclosure, there is provided a method for acquiring a content image, which is applied to a terminal, where an image pickup apparatus is disposed on the terminal, the method including:
at each preset time point after the content display process based on the target plane starts, acquiring a monitoring image containing display content on the target plane through the camera equipment;
determining whether the difference between two monitoring images acquired at two adjacent preset time points is greater than a preset difference threshold value or not;
under the condition that the difference degree is larger than the difference degree threshold value, generating a content image through a plurality of first monitoring images, wherein the content image comprises display content which is not shielded by a shielding object, and the first monitoring images are monitoring images acquired through the camera equipment after a target time point of which the difference degree is larger than the difference degree threshold value is determined;
after the content display process is determined to be terminated, a content image group corresponding to the content display process is obtained, wherein the content image group comprises a plurality of content images obtained before the content display process is terminated.
Optionally, the generating the content image by the plurality of first monitoring images includes:
determining whether each first monitoring image contains a shielding area, wherein the shielding area is an area which is shielded by the shielding object in an area corresponding to the display content in the first monitoring image;
under the condition that the plurality of first monitoring images contain the shielding areas, generating the content images according to the relative position information and the size information of the shielding areas in each first monitoring image; or,
and under the condition that a target monitoring image which does not contain the shielding area exists in the plurality of first monitoring images, taking the target monitoring image as the content image.
Optionally, the generating the content image according to the relative position information and the size information of the occlusion region in each of the first monitoring images includes:
determining a shielding area in each first monitoring image, the relative position information and the size information through a preset image recognition algorithm;
determining a reference image from the multiple first monitoring images according to the size information, wherein the size of a target shielding area contained in the reference image is smaller than a preset size;
determining a target image partition of the target shielding area in the reference image according to the relative position, wherein the target image partition is included in a plurality of image partitions obtained after the reference image is subjected to gridding processing;
extracting target subarea images from the plurality of first monitoring images, wherein the target subarea images are subarea images in a target image subarea of a second monitoring image in the plurality of first monitoring images, and the target image subarea of the second monitoring image does not contain a shielding area;
and replacing the subarea images in the target image subarea of the reference image by the target subarea image to acquire the content image.
Optionally, the determining whether a difference between two monitoring images acquired at two adjacent preset time points is greater than a preset difference threshold includes:
matching the two monitoring images acquired at two adjacent preset time points through a preset image matching algorithm to determine the difference degree;
comparing the difference to a difference threshold to determine whether the difference is greater than the difference threshold.
Optionally, after the acquiring of the content image group corresponding to the content display process, the method further includes:
and extracting keywords from the text content in each content image in the content image group to generate a content summary corresponding to each content image according to the extracted keywords.
According to a second aspect of the embodiments of the present disclosure, there is provided an apparatus for acquiring a content image, which is applied to a terminal on which an image pickup device is disposed, the apparatus including:
the image acquisition module is configured to acquire a monitoring image containing display content on a target plane through the camera equipment at each preset time point after the content display process based on the target plane starts;
the difference determining module is configured to determine whether the difference between the two monitoring images acquired at two adjacent preset time points is greater than a preset difference threshold value;
the image generation module is configured to generate a content image through a plurality of first monitoring images under the condition that the difference degree is determined to be larger than the difference degree threshold, wherein the content image comprises display content which is not shielded by a shielding object, and the first monitoring images are monitoring images acquired through the camera after a target time point of which the difference degree is larger than the difference degree threshold is determined;
the image group acquisition module is configured to acquire a content image group corresponding to the content presentation process after determining that the content presentation process is terminated, wherein the content image group comprises a plurality of content images acquired before the content presentation process is terminated.
Optionally, the image generating module is configured to:
determining whether each first monitoring image contains a shielding area, wherein the shielding area is an area which is shielded by the shielding object in an area corresponding to the display content in the first monitoring image;
under the condition that the plurality of first monitoring images contain the shielding areas, generating the content images according to the relative position information and the size information of the shielding areas in each first monitoring image; or,
and under the condition that a target monitoring image which does not contain the shielding area exists in the plurality of first monitoring images, taking the target monitoring image as the content image.
Optionally, the image generating module is configured to:
determining a shielding area in each first monitoring image, the relative position information and the size information through a preset image recognition algorithm;
determining a reference image from the multiple first monitoring images according to the size information, wherein the size of a target shielding area contained in the reference image is smaller than a preset size;
determining a target image partition of the target shielding area in the reference image according to the relative position, wherein the target image partition is included in a plurality of image partitions obtained after the reference image is subjected to gridding processing;
extracting target subarea images from the plurality of first monitoring images, wherein the target subarea images are subarea images in a target image subarea of a second monitoring image in the plurality of first monitoring images, and the target image subarea of the second monitoring image does not contain a shielding area;
and replacing the subarea images in the target image subarea of the reference image by the target subarea image to acquire the content image.
Optionally, the difference determining module is configured to:
matching the two monitoring images acquired at two adjacent preset time points through a preset image matching algorithm to determine the difference degree;
comparing the difference to a difference threshold to determine whether the difference is greater than the difference threshold.
Optionally, the apparatus further comprises:
the content summary generating module is configured to extract keywords from the text content in each content image in the content image group, so as to generate a content summary corresponding to each content image according to the extracted keywords.
According to a third aspect of embodiments of the present disclosure, there is provided a computer-readable storage medium having stored thereon computer program instructions, which when executed by a processor, implement the steps of the method of acquiring an image of content provided by the first aspect of the present disclosure.
According to a fourth aspect of the embodiments of the present disclosure, there is provided an electronic apparatus including: the second aspect of the disclosure provides an apparatus for acquiring a content image.
According to the technical scheme provided by the embodiment of the disclosure, at each preset time point after the content display process based on the target plane starts, the monitoring image containing the display content on the target plane can be acquired through the camera device; determining whether the difference between two adjacent monitoring images acquired at the preset time point is greater than a preset difference threshold value or not; under the condition that the difference degree is larger than the difference degree threshold value, generating a content image through a plurality of first monitoring images, wherein the content image comprises display content which is not shielded by a shielding object, and the first monitoring images are monitoring images acquired through the camera equipment after the target time point of which the difference degree is larger than the difference degree threshold value is determined; after the content presentation process is determined to be terminated, a content image group corresponding to the content presentation process is acquired, wherein the content image group comprises a plurality of content images acquired before the content presentation process is terminated. The change of the display content can be perceived in the content display process, and when the display content changes, a complete display content image is extracted from the collected monitoring images, so that the efficiency and the intelligent degree of the display content recording are improved.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the disclosure.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the present disclosure and together with the description, serve to explain the principles of the disclosure.
FIG. 1 is a flow diagram illustrating a method of obtaining an image of content in accordance with one illustrative embodiment;
FIG. 2 is a flow chart of a method of generating an image of content according to one embodiment shown in FIG. 1;
FIG. 3 is a flow chart of a method of determining the degree of difference of monitored images according to the method shown in FIG. 1;
FIG. 4 is a flow chart of another method of obtaining an image of content shown in FIG. 1;
FIG. 5 is a block diagram illustrating an apparatus for capturing images of content in accordance with one illustrative embodiment;
FIG. 6 is a block diagram of another apparatus for acquiring an image of content shown in FIG. 5;
FIG. 7 is a block diagram illustrating an electronic device in accordance with an example embodiment.
Detailed Description
Reference will now be made in detail to the exemplary embodiments, examples of which are illustrated in the accompanying drawings. When the following description refers to the accompanying drawings, like numbers in different drawings represent the same or similar elements unless otherwise indicated. The implementations described in the exemplary embodiments below are not intended to represent all implementations consistent with the present disclosure. Rather, they are merely examples of apparatus and methods consistent with certain aspects of the present disclosure, as detailed in the appended claims.
Before describing the method for acquiring a content image provided by the present disclosure, an application scenario related to various embodiments of the present disclosure is first described, where the application scenario includes a terminal, the terminal is provided with an image capturing device, and the terminal is arranged at a fixed position opposite to a target plane where a content is displayed. The terminal may be, for example, a mobile terminal such as a smart phone, a tablet computer, a smart watch, a smart bracelet, and a PDA (personal digital Assistant) provided with an infrared sensor and a camera device, or a fixed terminal such as a desktop computer and a smart television.
Fig. 1 is a flowchart illustrating a method for acquiring a content image according to an exemplary embodiment, and the method is applied to a terminal described in the application scenario, as shown in fig. 1, and includes the following steps:
in step 101, at each preset time point after the content presentation process based on the target plane is started, a monitoring image including the presentation content on the target plane is acquired by the image pickup apparatus.
Illustratively, the presentation may be a presentation or a blackboard. The target plane may be a plane where a display screen for playing a presentation screen is located, or a plane where a projector generates a presentation projection, or a plane where a blackboard of a writing board is located. The content display process may be a display process of a presentation or a writing process of a writing board. The triggering mode of the content display process comprises the following steps: image analysis or manual. Specifically, the image analysis method may include: continuously acquiring an image of a target plane through camera equipment of the terminal; the content presentation process may be determined to begin when the captured images show that the display screen or projector has been turned on (at which time the brightness of the target plane is greater than the preset brightness), or when the captured images show that the instructor has begun writing on a tablet (at which time the image recognition algorithm recognizes that there is a human action in the images). The manual mode may include: and after receiving a triggering instruction sent by a user, determining that the content presentation process starts.
For example, in the step 101, the image monitoring process for the target plane may be started at the same time when the content presentation process is determined to be started. Specifically, after determining that the content presentation process is started, the monitoring image of the presented content on the target plane may be collected once every preset time interval (every two intervals of the preset time point). The interval between every two preset time points can be set to be the same preset time length or different preset time lengths according to the actual requirement of the content display process. For example, the monitoring image is acquired every 5 seconds after the start of the content presentation process is determined, or the monitoring images are acquired every 3 rd second, 5 th second, 6 th second, 10 th second to nth second after the start of the content presentation process is determined.
In step 102, it is determined whether the difference between two of the monitored images acquired at two adjacent preset time points is greater than a preset difference threshold.
For example, two monitoring images acquired at two consecutive preset time points may be compared. If the difference between the two monitored images is too large (greater than the difference threshold), the instructor may be deemed to have updated the presentation, for example, the next presentation has been switched, or the instructor has erased the current blackboard writing and has rewritten another blackboard writing. If the difference between the two monitored images is small (less than or equal to the difference threshold), the presenter may be deemed to not have updated the presentation.
In step 103, in a case where it is determined that the degree of difference is larger than the degree of difference threshold value, a content image is generated from a plurality of first monitor images.
The content image comprises display content which is not shielded by a shielding object, and the first monitoring image is a monitoring image acquired by the camera equipment after the target time point with the difference degree larger than the difference degree threshold value is determined.
Illustratively, after determining that the instructor has updated the presentation, a content image may be generated through step 103 that includes the presentation and that does not have an obstruction in front of the presentation. The shelter may be the instructor itself or an instrument held by the instructor. It should be noted that, steps 101 to 103 actually need to be performed iteratively, and when it is determined that the difference between the two monitoring images is too large through this step 102, it is necessary to generate a content image for the display content in the monitoring image acquired later in step 103, and it can be considered that the content image of the display content in the monitoring image acquired earlier in the two monitoring images has been generated in the previous iteration process.
For example, a preset time period or generation of a content image may be used as a cutoff condition for acquiring the first detection image. Specifically, for the cutoff condition of the preset time duration, the plurality of first monitoring images may be acquired at preset time intervals within a preset time duration after the target time point. For the content image generation cutoff condition, the first monitoring image may be continuously acquired at preset time intervals after the target time point, and the content image may be determined according to the acquired first monitoring image until the content image is acquired.
In addition, it should be noted that after the content image is generated, the plurality of first monitoring images may be deleted from the storage space of the terminal, that is, after each iteration of step 101 to step 103 is performed, the terminal only stores the newly generated content image, so that the occupation of the storage space of the terminal by the method for acquiring the content image provided by the embodiment of the present disclosure may be reduced.
In step 104, after determining that the content presentation process is terminated, a content image group corresponding to the content presentation process is acquired.
Wherein the content image group includes a plurality of the content images acquired before the content presentation process is terminated.
Illustratively, before step 104, the above-mentioned steps 101 to 103 need to be iteratively executed until it is determined that the content presentation process is terminated, and the manner of determining that the content presentation process is terminated also includes: image analysis or manual. Specifically, the image analysis method herein may include: and when the difference degrees among the continuously acquired monitoring images in the preset number are smaller than a preset difference degree threshold value, determining that the content display process is terminated. The manual mode may include: and after receiving a termination instruction sent by the user, determining that the content presentation process is terminated. After determining that the content presentation process is terminated, a content image group including a plurality of content images generated throughout the content presentation process may be acquired. The plurality of content images in the content image group may be arranged and stored in the generated time sequence, or the page text in each content image may be identified by a preset text identification algorithm, and then the plurality of content images may be arranged and stored according to the identified page sequence. In addition, before the plurality of content images are combined into the content image group, each content image can be corrected through a preset image correction algorithm so as to improve the definition of each content image.
In summary, according to the technical solution provided by the embodiment of the present disclosure, at each preset time point after the content display process based on the target plane starts, the camera device can acquire the monitoring image including the display content on the target plane; determining whether the difference between two adjacent monitoring images acquired at the preset time point is greater than a preset difference threshold value or not; under the condition that the difference degree is larger than the difference degree threshold value, generating a content image through a plurality of first monitoring images, wherein the content image comprises display content which is not shielded by a shielding object, and the first monitoring images are monitoring images acquired through the camera equipment after the target time point of which the difference degree is larger than the difference degree threshold value is determined; after the content presentation process is determined to be terminated, a content image group corresponding to the content presentation process is acquired, wherein the content image group comprises a plurality of content images acquired before the content presentation process is terminated. The change of the display content can be perceived in the content display process, and when the display content changes, a complete display content image is extracted from the collected monitoring images, so that the efficiency and the intelligent degree of the display content recording are improved.
Fig. 2 is a flow chart of a method of generating a content image according to fig. 1, as shown in fig. 2, at step 103 may include: steps 1031 and 1032, or steps 1031 and 1033.
In step 1031, it is determined whether each of the first monitored images contains a masked region.
The shielding region is a region which is shielded by the shielding object in a region corresponding to the display content in the first monitoring image.
For example, the image in the mask region is clearly different from the image of the display content, for example, if the display content is a presentation, the brightness of the image of the display content is higher relative to the image in the mask region; and the display content is a blackboard-writing, the portrait can be detected in the image in the shielding area. Based on the above, whether the shading area exists in each first monitoring image can be determined by means of brightness recognition or portrait recognition aiming at different display contents.
In step 1032, in a case where it is determined that the plurality of first monitor images each include the mask region, the content image is generated based on the relative position information and the size information of the mask region in each of the first monitor images.
Illustratively, the shielding object is a person and/or a person-held instrument, and the shielding area is a shielding area formed by shielding the person and/or the person-held instrument between the display content and the image pickup apparatus. While the exposer may frequently move during the process of explaining the exhibition. Therefore, the shaded areas in the first monitoring images may be at different positions and have different sizes. If a shielding region exists in the partition B of the first monitoring image A and a shielding region does not exist in the partition B of the first monitoring image C, the partition image D in the partition B of the first monitoring image C can be extracted, and then the partition image D is filled to the partition B of the first monitoring image A, so that the shielding region in the first monitoring image A is removed. Specifically, this step 1032 may include: determining a shielding area in each first monitoring image, the relative position information and the size information through a preset image recognition algorithm; determining a reference image from the plurality of first monitoring images according to the size information, wherein the size of a target shielding area contained in the reference image is smaller than a preset size; determining a target image partition of the target shielding area in the reference image according to the relative position, wherein the target image partition is included in a plurality of image partitions acquired after the grid processing is carried out on the reference image; extracting a target subarea image from the plurality of first monitoring images, wherein the target subarea image is a subarea image in a target image subarea of a second monitoring image in the plurality of first monitoring images, and the target image subarea of the second monitoring image does not contain a shielding area; and replacing the subarea images in the target image subarea of the reference image by the target subarea image to acquire the content image.
In step 1033, when it is determined that there is a target monitor image that does not include the mask region among the plurality of first monitor images, the target monitor image is regarded as the content image.
For example, when it is determined that there is a target monitoring image that does not include the mask region among the plurality of first monitoring images, that is, when the first monitoring image that does not include the mask region is directly captured by the imaging device, the captured first monitoring image that does not include the mask region may be directly self-evaluated as the content image.
Fig. 3 is a flowchart of a method for determining the degree of difference of the monitored images according to fig. 1, and as shown in fig. 3, the step 102 may include:
in step 1021, the two monitoring images obtained at two adjacent preset time points are matched through a preset image matching algorithm to determine the difference.
In step 1022, the difference is compared with a difference threshold to determine whether the difference is greater than the difference threshold.
Fig. 4 is a flow chart of another method for obtaining an image of content shown in fig. 1, and after step 104, the method may further include:
in step 105, extracting keywords from the text content in each content image in the content image group to generate a summary corresponding to each content image according to the extracted keywords.
For example, the content image group is stored in a form of a picture, but existing picture display software cannot provide indexes of all content images like PPT application, and when the content image group includes a large number of pictures, a user has a high difficulty in searching for a content image, and therefore, a corresponding content summary can be generated for each content image in step 105, so that the difficulty in searching for a content image by the user is reduced, and the searching efficiency is improved. Specifically, for each content image, all texts in the image can be identified through a preset text identification algorithm, one or more keywords are extracted from all texts in the image through a preset keyword extraction model, and then the one or more keywords are combined to generate the content summary. The content digest is set as the file name of the content image or the content digest is saved in the attribute information of the content image.
In another embodiment of the present disclosure, while the content images are generated through the above step 103, time period information corresponding to each content image may be recorded, where the time period information is used to record a time period for which the display content corresponding to the content image keeps the display state. And audio information sent by the instructor is collected during the content display process. The audio information also corresponds to different points in time. And combining the audio information and the time period of the display content corresponding to the content image to keep the display state, namely acquiring an entire audio segment in the content display process to extract the audio segment corresponding to each content image, and storing the audio segment corresponding to each content image.
In summary, according to the technical solution provided by the embodiment of the present disclosure, at each preset time point after the content display process based on the target plane starts, the camera device can acquire the monitoring image including the display content on the target plane; determining whether the difference between two adjacent monitoring images acquired at the preset time point is greater than a preset difference threshold value or not; under the condition that the difference degree is larger than the difference degree threshold value, generating a content image through a plurality of first monitoring images, wherein the content image comprises display content which is not shielded by a shielding object, and the first monitoring images are monitoring images acquired through the camera equipment after the target time point of which the difference degree is larger than the difference degree threshold value is determined; after the content presentation process is determined to be terminated, a content image group corresponding to the content presentation process is acquired, wherein the content image group comprises a plurality of content images acquired before the content presentation process is terminated. The change of the display content can be perceived in the content display process, and when the display content changes, a complete display content image is extracted from the collected monitoring images, so that the efficiency and the intelligent degree of the display content recording are improved.
Fig. 5 is a block diagram illustrating an apparatus for acquiring a content image according to an exemplary embodiment, and as shown in fig. 5, the apparatus 500 is applied to a terminal of the application scenario, and includes:
an image obtaining module 510, configured to obtain, at each preset time point after a content display process based on a target plane starts, a monitor image including display content on the target plane through the image capturing apparatus;
a difference determining module 520 configured to determine whether a difference between two adjacent monitored images acquired at the preset time point is greater than a preset difference threshold;
an image generating module 530, configured to, in a case that it is determined that the difference degree is greater than the difference degree threshold, generate a content image through a plurality of first monitoring images, where the content image includes display content that is not blocked by a shielding object, and the first monitoring images are monitoring images acquired by the image capturing apparatus after determining a target time point at which the difference degree is greater than the difference degree threshold;
an image group acquiring module 540, configured to, after determining that the content presentation process is terminated, acquire a content image group corresponding to the content presentation process, where the content image group includes a plurality of content images acquired before the content presentation process is terminated.
Optionally, the image generating module 530 is configured to:
determining whether each first monitoring image contains a shielding area, wherein the shielding area is an area which is shielded by the shielding object in an area corresponding to the display content in the first monitoring image;
under the condition that the plurality of first monitoring images are determined to contain the shielding area, generating the content image according to the relative position information and the size information of the shielding area in each first monitoring image; or,
and when determining that the target monitoring image which does not contain the shielding area exists in the plurality of first monitoring images, taking the target monitoring image as the content image.
Optionally, the image generating module 530 is configured to:
determining a shielding area in each first monitoring image, the relative position information and the size information through a preset image recognition algorithm;
determining a reference image from the plurality of first monitoring images according to the size information, wherein the size of a target shielding area contained in the reference image is smaller than a preset size;
determining a target image partition of the target shielding area in the reference image according to the relative position, wherein the target image partition is included in a plurality of image partitions acquired after the grid processing is carried out on the reference image;
extracting a target subarea image from the plurality of first monitoring images, wherein the target subarea image is a subarea image in a target image subarea of a second monitoring image in the plurality of first monitoring images, and the target image subarea of the second monitoring image does not contain a shielding area;
and replacing the subarea images in the target image subarea of the reference image by the target subarea image to acquire the content image.
Optionally, the difference determining module 520 is configured to:
matching the two monitoring images acquired at two adjacent preset time points through a preset image matching algorithm to determine the difference degree;
the difference is compared to a difference threshold to determine whether the difference is greater than the difference threshold.
Fig. 6 is a block diagram of another apparatus for acquiring an image of contents shown in fig. 5, and as shown in fig. 7, the apparatus 500 further includes:
a content summary generating module 550 configured to perform keyword extraction on the text content in each content image in the content image group, so as to generate a content summary corresponding to each content image according to the extracted keyword.
In summary, according to the technical solution provided by the embodiment of the present disclosure, at each preset time point after the content display process based on the target plane starts, the camera device can acquire the monitoring image including the display content on the target plane; determining whether the difference between two adjacent monitoring images acquired at the preset time point is greater than a preset difference threshold value or not; under the condition that the difference degree is larger than the difference degree threshold value, generating a content image through a plurality of first monitoring images, wherein the content image comprises display content which is not shielded by a shielding object, and the first monitoring images are monitoring images acquired through the camera equipment after the target time point of which the difference degree is larger than the difference degree threshold value is determined; after the content presentation process is determined to be terminated, a content image group corresponding to the content presentation process is acquired, wherein the content image group comprises a plurality of content images acquired before the content presentation process is terminated. The change of the display content can be perceived in the content display process, and when the display content changes, a complete display content image is extracted from the collected monitoring images, so that the efficiency and the intelligent degree of the display content recording are improved.
FIG. 7 is a block diagram illustrating an electronic device in accordance with an example embodiment. For example, the electronic device 700 may be, for example, a mobile telephone, a computer, a digital broadcast terminal, a messaging device, a game console, a tablet device, a medical device, an exercise device, a personal digital assistant, or the like.
Referring to fig. 7, electronic device 700 may include one or more of the following components: a processing component 702, a memory 704, a power component 706, a multimedia component 707, an audio component 710, an input/output (I/O) interface 712, a sensor component 714, and a communication component 716.
The processing component 702 generally controls overall operation of the electronic device 700, such as operations associated with display, telephone calls, data communications, camera operations, and recording operations. The processing component 702 may include one or more processors 720 to execute instructions to perform all or a portion of the steps of the method of obtaining an image of content described above. Further, the processing component 702 may include one or more modules that facilitate interaction between the processing component 702 and other components. For example, the processing component 702 may include a multimedia module to facilitate interaction between the multimedia component 707 and the processing component 702.
The memory 704 is configured to store various types of data to support operations at the electronic device 700. Examples of such data include instructions for any application or method operating on the electronic device 700, contact data, phonebook data, messages, pictures, videos, and so forth. The memory 704 may be implemented by any type or combination of volatile or non-volatile memory devices such as Static Random Access Memory (SRAM), electrically erasable programmable read-only memory (EEPROM), erasable programmable read-only memory (EPROM), programmable read-only memory (PROM), read-only memory (ROM), magnetic memory, flash memory, magnetic or optical disks.
The power component 706 provides power to the various components of the electronic device 700. The power components 706 may include a power management system, one or more power sources, and other components associated with generating, managing, and distributing power for the electronic device 700.
The multimedia component 707 includes a screen providing an output interface between the electronic device 700 and a user. In some embodiments, the screen may include a Liquid Crystal Display (LCD) and a Touch Panel (TP). If the screen includes a touch panel, the screen may be implemented as a touch screen to receive an input signal from a user. The touch panel includes one or more touch sensors to sense touch, slide, and gestures on the touch panel. The touch sensor may not only sense the boundary of a touch or slide action, but also detect the duration and pressure associated with the touch or slide operation. In some embodiments, the multimedia component 707 includes a front facing camera and/or a rear facing camera. The front camera and/or the rear camera may receive external multimedia data when the electronic device 700 is in an operation mode, such as a photographing mode or a video mode. Each front camera and rear camera may be a fixed optical lens system or have a focal length and optical zoom capability.
The audio component 710 is configured to output and/or input audio signals. For example, the audio component 710 includes a Microphone (MIC) configured to receive external audio signals when the electronic device 700 is in an operational mode, such as a call mode, a recording mode, and a voice recognition mode. The received audio signal may further be stored in the memory 704 or transmitted via the communication component 716. In some embodiments, audio component 710 also includes a speaker for outputting audio signals.
The I/O interface 712 provides an interface between the processing component 702 and peripheral interface modules, which may be keyboards, click wheels, buttons, etc. These buttons may include, but are not limited to: a home button, a volume button, a start button, and a lock button.
The sensor assembly 714 includes one or more sensors for providing various aspects of status assessment for the electronic device 700. For example, the sensor assembly 714 may detect an open/closed state of the electronic device 700, the relative positioning of components, such as a display and keypad of the electronic device 700, the sensor assembly 714 may also detect a change in the position of the electronic device 700 or a component of the electronic device 700, the presence or absence of user contact with the electronic device 700, orientation or acceleration/deceleration of the electronic device 700, and a change in the temperature of the electronic device 700. The sensor assembly 714 may include a proximity sensor configured to detect the presence of a nearby object without any physical contact. The sensor assembly 714 may also include a light sensor, such as a CMOS or CCD image sensor, for use in imaging applications. In some embodiments, the sensor assembly 714 may also include an acceleration sensor, a gyroscope sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
The communication component 716 is configured to facilitate wired or wireless communication between the electronic device 700 and other devices. The electronic device 700 may access a wireless network based on a communication standard, such as WiFi, 2G or 3G, or a combination thereof. In an exemplary embodiment, the communication component 716 receives a broadcast signal or broadcast related information from an external broadcast management system via a broadcast channel. In an exemplary embodiment, the communication component 716 further includes a Near Field Communication (NFC) module to facilitate short-range communications. For example, the NFC module may be implemented based on Radio Frequency Identification (RFID) technology, infrared data association (IrDA) technology, Ultra Wideband (UWB) technology, Bluetooth (BT) technology, and other technologies.
In an exemplary embodiment, the electronic device 700 may be implemented by one or more Application Specific Integrated Circuits (ASICs), Digital Signal Processors (DSPs), Digital Signal Processing Devices (DSPDs), Programmable Logic Devices (PLDs), Field Programmable Gate Arrays (FPGAs), controllers, micro-controllers, microprocessors or other electronic components for performing the above-described method of acquiring content images.
In an exemplary embodiment, a non-transitory computer-readable storage medium comprising instructions, such as the memory 704 comprising instructions, executable by the processor 720 of the electronic device 700 to perform the above-described method of capturing a content image is also provided. For example, the non-transitory computer readable storage medium may be a ROM, a Random Access Memory (RAM), a CD-ROM, a magnetic tape, a floppy disk, an optical data storage device, and the like.
In another exemplary embodiment, a computer program product is also provided, which comprises a computer program executable by a programmable apparatus, the computer program having code portions for performing the above-mentioned method of acquiring an image of content when executed by the programmable apparatus.
To sum up, this disclosure can be in the change of content show in-process perception show content to when show content appears changing from the complete show content image of drawing in a plurality of monitoring images of gathering, improve the efficiency and the intelligent degree that show content recorded.
Other embodiments of the disclosure will be apparent to those skilled in the art from consideration of the specification and practice of the disclosure. This application is intended to cover any variations, uses, or adaptations of the disclosure following, in general, the principles of the disclosure and including such departures from the present disclosure as come within known or customary practice within the art to which the disclosure pertains. It is intended that the specification and examples be considered as exemplary only, with a true scope and spirit of the disclosure being indicated by the following claims.
It will be understood that the present disclosure is not limited to the precise arrangements described above and shown in the drawings and that various modifications and changes may be made without departing from the scope thereof. The scope of the present disclosure is limited only by the appended claims.
Claims (12)
1. A method for acquiring content images is applied to a terminal, wherein a camera device is arranged on the terminal, and the method comprises the following steps:
at each preset time point after the content display process based on the target plane starts, acquiring a monitoring image containing display content on the target plane through the camera equipment;
determining whether the difference between two monitoring images acquired at two adjacent preset time points is greater than a preset difference threshold value or not;
under the condition that the difference degree is larger than the difference degree threshold value, generating a content image through a plurality of first monitoring images, wherein the content image comprises display content which is not shielded by a shielding object, and the first monitoring images are monitoring images acquired through the camera equipment after a target time point of which the difference degree is larger than the difference degree threshold value is determined;
after the content display process is determined to be terminated, a content image group corresponding to the content display process is obtained, wherein the content image group comprises a plurality of content images obtained before the content display process is terminated.
2. The method of claim 1, wherein generating the content image from the plurality of first monitored images comprises:
determining whether each first monitoring image contains a shielding area, wherein the shielding area is an area which is shielded by the shielding object in an area corresponding to the display content in the first monitoring image;
under the condition that the plurality of first monitoring images contain the shielding areas, generating the content images according to the relative position information and the size information of the shielding areas in each first monitoring image; or,
and under the condition that a target monitoring image which does not contain the shielding area exists in the plurality of first monitoring images, taking the target monitoring image as the content image.
3. The method of claim 2, wherein generating the content image according to the relative position information and the size information of the occlusion region in each of the first monitored images comprises:
determining a shielding area in each first monitoring image, the relative position information and the size information through a preset image recognition algorithm;
determining a reference image from the multiple first monitoring images according to the size information, wherein the size of a target shielding area contained in the reference image is smaller than a preset size;
determining a target image partition of the target shielding area in the reference image according to the relative position, wherein the target image partition is included in a plurality of image partitions obtained after the reference image is subjected to gridding processing;
extracting target subarea images from the plurality of first monitoring images, wherein the target subarea images are subarea images in a target image subarea of a second monitoring image in the plurality of first monitoring images, and the target image subarea of the second monitoring image does not contain a shielding area;
and replacing the subarea images in the target image subarea of the reference image by the target subarea image to acquire the content image.
4. The method according to claim 1, wherein the determining whether the difference between two of the monitored images acquired at two adjacent preset time points is greater than a preset difference threshold value comprises:
matching the two monitoring images acquired at two adjacent preset time points through a preset image matching algorithm to determine the difference degree;
comparing the difference to a difference threshold to determine whether the difference is greater than the difference threshold.
5. The method according to claim 1, wherein after the acquiring of the group of content images corresponding to the content presentation process, the method further comprises:
and extracting keywords from the text content in each content image in the content image group to generate a content summary corresponding to each content image according to the extracted keywords.
6. An apparatus for obtaining content images, applied to a terminal, on which a camera device is disposed, the apparatus comprising:
the image acquisition module is configured to acquire a monitoring image containing display content on a target plane through the camera equipment at each preset time point after the content display process based on the target plane starts;
the difference determining module is configured to determine whether the difference between the two monitoring images acquired at two adjacent preset time points is greater than a preset difference threshold value;
the image generation module is configured to generate a content image through a plurality of first monitoring images under the condition that the difference degree is determined to be larger than the difference degree threshold, wherein the content image comprises display content which is not shielded by a shielding object, and the first monitoring images are monitoring images acquired through the camera after a target time point of which the difference degree is larger than the difference degree threshold is determined;
the image group acquisition module is configured to acquire a content image group corresponding to the content presentation process after determining that the content presentation process is terminated, wherein the content image group comprises a plurality of content images acquired before the content presentation process is terminated.
7. The apparatus of claim 6, wherein the image generation module is configured to:
determining whether each first monitoring image contains a shielding area, wherein the shielding area is an area which is shielded by the shielding object in an area corresponding to the display content in the first monitoring image;
under the condition that the plurality of first monitoring images contain the shielding areas, generating the content images according to the relative position information and the size information of the shielding areas in each first monitoring image; or,
and under the condition that a target monitoring image which does not contain the shielding area exists in the plurality of first monitoring images, taking the target monitoring image as the content image.
8. The apparatus of claim 7, wherein the image generation module is configured to:
determining a shielding area in each first monitoring image, the relative position information and the size information through a preset image recognition algorithm;
determining a reference image from the multiple first monitoring images according to the size information, wherein the size of a target shielding area contained in the reference image is smaller than a preset size;
determining a target image partition of the target shielding area in the reference image according to the relative position, wherein the target image partition is included in a plurality of image partitions obtained after the reference image is subjected to gridding processing;
extracting target subarea images from the plurality of first monitoring images, wherein the target subarea images are subarea images in a target image subarea of a second monitoring image in the plurality of first monitoring images, and the target image subarea of the second monitoring image does not contain a shielding area;
and replacing the subarea images in the target image subarea of the reference image by the target subarea image to acquire the content image.
9. The apparatus of claim 6, wherein the disparity determination module is configured to:
matching the two monitoring images acquired at two adjacent preset time points through a preset image matching algorithm to determine the difference degree;
comparing the difference to a difference threshold to determine whether the difference is greater than the difference threshold.
10. The apparatus of claim 6, further comprising:
the content summary generating module is configured to extract keywords from the text content in each content image in the content image group, so as to generate a content summary corresponding to each content image according to the extracted keywords.
11. A computer-readable storage medium, on which computer program instructions are stored, which program instructions, when executed by a processor, carry out the steps of the method according to any one of claims 1 to 5.
12. An electronic device, comprising: apparatus for capturing an image of content as claimed in any of claims 6 to 10.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010622912.8A CN111832455A (en) | 2020-06-30 | 2020-06-30 | Method, device, storage medium and electronic equipment for acquiring content image |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010622912.8A CN111832455A (en) | 2020-06-30 | 2020-06-30 | Method, device, storage medium and electronic equipment for acquiring content image |
Publications (1)
Publication Number | Publication Date |
---|---|
CN111832455A true CN111832455A (en) | 2020-10-27 |
Family
ID=72900740
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010622912.8A Pending CN111832455A (en) | 2020-06-30 | 2020-06-30 | Method, device, storage medium and electronic equipment for acquiring content image |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111832455A (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113873292A (en) * | 2021-09-30 | 2021-12-31 | 联想(北京)有限公司 | Processing method and electronic equipment |
CN114694434A (en) * | 2020-12-28 | 2022-07-01 | 康立 | Video teaching course intelligent generation method and system based on deep learning |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160021281A1 (en) * | 2014-07-18 | 2016-01-21 | Xiaomi Inc. | Photographing method and device |
CN109167974A (en) * | 2018-10-25 | 2019-01-08 | 洪磊 | A kind of writing on the blackboard recording system |
CN109492206A (en) * | 2018-10-10 | 2019-03-19 | 深圳市容会科技有限公司 | PPT presentation file method for recording, device, computer equipment and storage medium |
CN111104913A (en) * | 2019-12-23 | 2020-05-05 | 福州大学 | Video PPT extraction method based on structure and similarity |
-
2020
- 2020-06-30 CN CN202010622912.8A patent/CN111832455A/en active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160021281A1 (en) * | 2014-07-18 | 2016-01-21 | Xiaomi Inc. | Photographing method and device |
CN109492206A (en) * | 2018-10-10 | 2019-03-19 | 深圳市容会科技有限公司 | PPT presentation file method for recording, device, computer equipment and storage medium |
CN109167974A (en) * | 2018-10-25 | 2019-01-08 | 洪磊 | A kind of writing on the blackboard recording system |
CN111104913A (en) * | 2019-12-23 | 2020-05-05 | 福州大学 | Video PPT extraction method based on structure and similarity |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114694434A (en) * | 2020-12-28 | 2022-07-01 | 康立 | Video teaching course intelligent generation method and system based on deep learning |
CN114694434B (en) * | 2020-12-28 | 2023-12-01 | 康立 | Video teaching course intelligent generation method and system based on deep learning |
CN113873292A (en) * | 2021-09-30 | 2021-12-31 | 联想(北京)有限公司 | Processing method and electronic equipment |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN106651955B (en) | Method and device for positioning target object in picture | |
EP3182716A1 (en) | Method and device for video display | |
CN105845124B (en) | Audio processing method and device | |
EP3179408A2 (en) | Picture processing method and apparatus, computer program and recording medium | |
US11094110B2 (en) | Method, apparatus and electronic device for processing image | |
EP3176709A1 (en) | Video categorization method and apparatus, computer program and recording medium | |
WO2016192325A1 (en) | Method and device for processing logo on video file | |
CN106791535B (en) | Video recording method and device | |
CN106534951B (en) | Video segmentation method and device | |
CN112991553B (en) | Information display method and device, electronic equipment and storage medium | |
EP3796317A1 (en) | Video processing method, video playing method, devices and storage medium | |
EP3147802B1 (en) | Method and apparatus for processing information | |
WO2017080084A1 (en) | Font addition method and apparatus | |
CN110796094A (en) | Control method and device based on image recognition, electronic equipment and storage medium | |
CN109766473B (en) | Information interaction method and device, electronic equipment and storage medium | |
US20220222831A1 (en) | Method for processing images and electronic device therefor | |
CN106331328B (en) | Information prompting method and device | |
CN106254939B (en) | Information prompting method and device | |
CN104850643B (en) | Picture comparison method and device | |
CN113032627A (en) | Video classification method and device, storage medium and terminal equipment | |
CN107885571B (en) | Display page control method and device | |
CN110968364A (en) | Method and device for adding shortcut plug-in and intelligent equipment | |
CN111832455A (en) | Method, device, storage medium and electronic equipment for acquiring content image | |
CN109145878B (en) | Image extraction method and device | |
CN107105311B (en) | Live broadcasting method and device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |