CN113630546A - Image processing method, device and system - Google Patents

Image processing method, device and system Download PDF

Info

Publication number
CN113630546A
CN113630546A CN202010389013.8A CN202010389013A CN113630546A CN 113630546 A CN113630546 A CN 113630546A CN 202010389013 A CN202010389013 A CN 202010389013A CN 113630546 A CN113630546 A CN 113630546A
Authority
CN
China
Prior art keywords
image data
image
display screen
pixel
sensor
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010389013.8A
Other languages
Chinese (zh)
Inventor
闫玮
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Alibaba Group Holding Ltd
Original Assignee
Alibaba Group Holding Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Alibaba Group Holding Ltd filed Critical Alibaba Group Holding Ltd
Priority to CN202010389013.8A priority Critical patent/CN113630546A/en
Publication of CN113630546A publication Critical patent/CN113630546A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/80Camera processing pipelines; Components thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/61Control of cameras or camera modules based on recognised objects
    • H04N23/611Control of cameras or camera modules based on recognised objects where the recognised objects include parts of the human body
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/67Focus control based on electronic image sensor signals

Abstract

An image processing method, apparatus and system are disclosed, which process an image sensed by an image sensor to be displayed on a display screen. The number of pixels M × N of the image sensor is larger than the number of pixels R × S to be supplied to the display screen. First image data corresponding to the first area, the number of pixels m1 × n1 of the first image data, is read from the image sensor. The first image data is converted into second image data, and the number of pixels of the second image data is R multiplied by S. And sending the second image data to the display screen. In response to a zoom instruction, third image data corresponding to the second area is read from the image sensor, the number of pixels of the third image data being m2 × n2, m2 ≠ m1, n2 ≠ n 1. The third image data is converted into fourth image data having a pixel number of R × S. And sending the fourth image data to the display screen. Therefore, a clear zooming display effect on the display screen is conveniently realized without additionally increasing the hardware cost.

Description

Image processing method, device and system
Technical Field
The present disclosure relates to the field of image processing, and more particularly, to an image processing scheme for processing an image sensed by an image sensor for display on a display screen.
Background
With the development of smart home technology, some devices add a home scene real-time viewing function, and images shot by a camera arranged in a home on site can be presented on a display screen arranged in the home on site, and can also be sent to electronic devices, such as smart phones, through a network so as to be presented on the electronic devices.
The depth of field of the camera used in such a scene is often large, and for example, objects in a range of tens of centimeters to tens of meters can be clearly imaged. The image sensor used also tends to have a higher resolution, being able to sense the acquisition of very sharp images.
In addition, in order to present a real-time image of a wider angle of view range, a wide-angle camera is generally employed. In this way, a large range of real-time images can be acquired.
However, for objects at a distance from the camera, the image is relatively small in the entire real-time image presented to the user on the display screen. At this time, the detail expressiveness of the displayed object image is poor.
Even if the camera supports high resolution, the image sensor collects high-definition images, and for a long-distance object, only a small object image can be presented on the display screen. For example, the image of the object of interest may be one tenth or even less of the entire real-time image presented on the display screen. In this case, the user experience is greatly affected.
In this case, if the local zoom can be realized, the image of the object of interest is enlarged, which brings a better user experience to the client.
The conventional zoom method includes a digital zoom method and an optical zoom method.
The digital zooming mode cannot improve the detail expression capability of the object.
Optical zooming has a better effect, but optical zooming requires high optical costs. For example, a lens apparatus having an optical zoom capability needs to be attached.
Therefore, there is still a need for an image processing scheme that can conveniently achieve zooming.
Disclosure of Invention
One technical problem to be solved by the present disclosure is to provide an image processing scheme that can conveniently achieve a clear zoom effect without increasing additional hardware investment cost.
According to a first aspect of the present disclosure, there is provided an image processing method for processing an image sensed by an image sensor to be displayed on a display screen, wherein the number of pixels of sensor image data sensed by the image sensor is mxn, the number of pixels of image data supplied to the display screen or a device to which the display screen belongs is rxs, R and M are the number of pixel rows, respectively, and R < M, S and N are the number of pixel columns, respectively, and S < N, the method comprising: reading first image data corresponding to the first area from the image sensor, wherein the number of pixels of the first image data is M1 multiplied by N1, M1 is the number of pixel rows, R is not less than M1 and not more than M, N1 is the number of pixel columns, and S is not less than N1 and not more than N; converting the first image data into second image data, wherein the pixel number of the second image data is R multiplied by S; sending the second image data to a display screen or equipment to which the display screen belongs so as to display images; reading third image data corresponding to the second area from the image sensor in response to a zoom instruction, the third image data having a pixel count of M2 × N2, M2 being a pixel row number, and R ≦ M2 ≦ M, N2 being a pixel column number, and S ≦ N2 ≦ N, M2 ≠ M1, N2 ≠ N1; converting the third image data into fourth image data, wherein the pixel number of the fourth image data is R multiplied by S; and sending the fourth image data to the display screen or the equipment to which the display screen belongs so as to display the image.
According to a second aspect of the present disclosure, there is provided an image processing method for processing an image sensed by an image sensor to be displayed on a display screen, wherein the number of pixels of sensor image data sensed by the image sensor is mxn, the number of pixels of image data supplied to the display screen or a device to which the display screen belongs is rxs, R and M are the number of pixel rows, respectively, and R < M, S and N are the number of pixel columns, respectively, and S < N, the method comprising: reading sensor image data from an image sensor and storing the sensor image data in a memory; extracting first image data corresponding to the first region from the sensor image data on the memory, wherein the number of pixels of the first image data is M1 Xn 1, M1 is the number of pixel rows, R is not less than M1 and not more than M, N1 is the number of pixel columns, and S is not less than N1 and not more than N; converting the first image data into second image data, wherein the pixel number of the second image data is R multiplied by S; sending the second image data to a display screen or equipment to which the display screen belongs so as to display images; in response to a zoom instruction, extracting third image data corresponding to the second area from the sensor image data on the memory, the third image data having a number of pixels of M2 × N2, M2 being a number of pixel rows, and R ≦ M2 ≦ M, N2 being a number of pixel columns, and S ≦ N2 ≦ N, M2 ≠ M1, N2 ≠ N1; converting the third image data into fourth image data, wherein the pixel number of the fourth image data is R multiplied by S; and sending the fourth image data to the display screen or the equipment to which the display screen belongs so as to display the image.
Optionally, the center of the second region coincides with the center of the first region; or the second area is a designated area in the first area or an area where the designated object is located.
Optionally, the method may further include: the center of the second area is determined according to the zoom instruction.
Optionally, the method may further include: and receiving a zooming instruction from the equipment and/or the server to which the display screen belongs.
Optionally, the method may further include: identifying a specified object from the second image data; in response to identifying the specified object in the second image data, a zoom instruction is issued, the second region being a region in which the specified object is located.
Optionally, the method may further include: the second area is moved along with the movement of the specified object.
Optionally, the method may further include: in response to a partial or full movement of the second region beyond the range of the first region, the range of the first region on the image sensor is expanded and/or moved such that the expanded and/or moved first region includes the second region.
Optionally, the method may further include: and in response to the part or the whole of the second area moving beyond the range of the first area, rotating the shooting direction of the imaging device to which the image sensor belongs, so that the first area on the rotated image sensor comprises the second area.
Alternatively, the specified object is a face image, and the specified object is recognized from the second image data using a face recognition function.
Optionally, the method may further include: the face recognition function is turned on or off in response to the user's setting.
Optionally, the step of issuing a zoom instruction in response to identifying the specified object in the second image data comprises: when the distance from the specified object to the imaging device to which the image sensor belongs is greater than or equal to a first distance threshold value, sending a zooming instruction; and/or not issuing a zoom instruction in the case that the distance from the specified object to the imaging device to which the image sensor belongs is less than the first distance threshold.
Optionally, the method may further include: after the fourth image data is transmitted, in response to the distance from the specified object to the imaging device to which the image sensor belongs becoming smaller than the second distance threshold, the second image data is retransmitted to the display screen or the device to which the display screen belongs for image display.
Optionally, the method may further include: the number of pixels m2 × n2 of the third image data is determined based on the screen aspect ratio of the specified object with respect to the first area and the number of pixels m1 × n1 of the first image data.
Optionally, the step of issuing a zoom instruction in response to identifying the specified object in the second image data comprises: when the frame occupation ratio of the designated object relative to the first area is less than or equal to a first occupation ratio threshold value, sending a zooming instruction; and/or not issuing a zoom instruction when the screen occupation ratio of the designated object relative to the first area is greater than a first occupation ratio threshold.
Optionally, the method may further include: after the fourth image data is transmitted, in response to the screen aspect ratio of the specified object with respect to the first area becoming larger than the second aspect ratio threshold, the second image data is retransmitted to the display screen or the device to which the display screen belongs for image display.
Optionally, the method may further include: based on the fourth image data, the identity information of the specified object is recognized using a face recognition function.
Optionally, the method may further include: based on the identified identity information, granting the authority of the user corresponding to the specified object to execute the predetermined operation; and/or performing security check on the user corresponding to the specified object based on the identified identity information.
Optionally, the second area is a part of the first area, and further includes, before sending the fourth image data to the display screen or the device to which the display screen belongs for image display in response to the zoom instruction: a dynamic process from displaying the image corresponding to the first area to displaying the image corresponding to the second area is presented on the display screen by a digital zoom process based on the second image data.
Optionally, the method may further include: and responding to the display screen or the equipment to which the display screen belongs to receive the fourth image data, ending the dynamic change process display, and displaying the image based on the fourth image data.
Optionally, the display screen is provided on the same device as the image sensor.
Optionally, the display screen and the image sensor are disposed on different devices, and the second image data and the fourth image data are sent to the device to which the display screen belongs through a network.
Optionally, the method may further include: receiving an image from a device to which a display screen belongs; the received image is displayed on a local display screen.
Optionally, the image sensor senses live images in real time.
According to a third aspect of the present disclosure, there is provided an image processing method including: identifying a specified object from the image data in response to the interactive request; and executing zoom processing to increase the screen occupation ratio of the designated object in the image data after the zoom processing in a case where the screen occupation ratio of the designated object in the image data is smaller than a predetermined occupation ratio threshold.
Alternatively, the specified object is a face image, and the specified object is recognized from the second image data using a face recognition function.
Optionally, the method may further include: the face recognition function is turned on or off in response to the user's setting.
According to a fourth aspect of the present disclosure, there is provided an image processing method for processing an image sensed by an image sensor to be displayed on a display screen, wherein the number of pixels of sensor image data sensed by the image sensor is mxn, the number of pixels of image data supplied to the display screen is rxs, R and M are the number of pixel rows, respectively, and R < M, S and N are the number of pixel columns, respectively, and S < N. The method comprises the following steps: performing image conversion to obtain first image data corresponding to a first area on the image sensor and second image data for displaying on a display screen, wherein the pixel number of the first image data is M1 × N1, M1 is the pixel row number, R is not less than M1 and not more than M, N1 is the pixel column number, S is not less than N1 and not more than N, and the pixel number of the second image data is R × S; when the distance from the specified object to the imaging device to which the image sensor belongs is greater than or equal to a first distance threshold value, sending a zooming instruction; and performing image conversion in response to a zoom instruction to obtain third image data corresponding to a second area on the image sensor and fourth image data for display on the display screen, wherein the second area is an area where the designated object is located, the number of pixels of the third image data is M2 × N2, M2 is the number of pixel rows, R is not less than M2 and not more than M, N2 is the number of pixel columns, S is not less than N2 and not more than N, M2 is not equal to M1, N2 is not equal to N1, and the number of pixels of the fourth image data is R × S.
Optionally, the method may further include: after the fourth image data is transmitted to the display screen or the device to which the display screen belongs for image display, in response to the distance from the specified object to the imaging device to which the image sensor belongs becoming smaller than the second distance threshold, the second image data is retransmitted to the display screen or the device to which the display screen belongs for image display.
According to a fifth aspect of the present disclosure, there is provided an image processing method for processing an image sensed by an image sensor to be displayed on a display screen, wherein the number of pixels of sensor image data sensed by the image sensor is M × N, the number of pixels of image data supplied to the display screen is R × S, R and M are the number of pixel rows, respectively, and R < M, S and N are the number of pixel columns, respectively, and S < N. The method comprises the following steps: performing image conversion to obtain first image data corresponding to a first area on the image sensor and second image data for displaying on a display screen, wherein the pixel number of the first image data is M1 × N1, M1 is the pixel row number, R is not less than M1 and not more than M, N1 is the pixel column number, S is not less than N1 and not more than N, and the pixel number of the second image data is R × S; when the frame occupation ratio of the designated object relative to the first area is less than or equal to a first occupation ratio threshold value, sending a zooming instruction; and performing image conversion in response to a zoom instruction to obtain third image data corresponding to a second area on the image sensor and fourth image data for display on the display screen, wherein the second area is an area where the designated object is located, the number of pixels of the third image data is M2 × N2, M2 is the number of pixel rows, R is not less than M2 and not more than M, N2 is the number of pixel columns, S is not less than N2 and not more than N, M2 is not equal to M1, N2 is not equal to N1, and the number of pixels of the fourth image data is R × S.
Optionally, the method may further include: after the fourth image data is transmitted to the display screen or the device to which the display screen belongs for image display, in response to the screen occupation ratio of the specified object with respect to the first area becoming larger than the second occupation ratio threshold, the second image data is transmitted again to the display screen or the device to which the display screen belongs for image display.
Optionally, the method may further include: the number of pixels m2 × n2 of the third image data is determined based on the screen aspect ratio of the specified object with respect to the first area and the number of pixels m1 × n1 of the first image data.
According to a sixth aspect of the present disclosure, there is provided an image processing method for processing an image sensed by an image sensor to be displayed on a display screen, wherein the number of pixels of sensor image data sensed by the image sensor is mxn, the number of pixels of image data supplied to the display screen is rxs, R and M are the number of pixel rows, respectively, and R < M, S and N are the number of pixel columns, respectively, and S < N, the method comprising: performing image conversion to obtain first image data corresponding to a first area on the image sensor and second image data for displaying on a display screen, wherein the pixel number of the first image data is M1 × N1, M1 is the pixel row number, R is not less than M1 and not more than M, N1 is the pixel column number, S is not less than N1 and not more than N, and the pixel number of the second image data is R × S; identifying a specified object from the second image data; responding to a zoom instruction, performing image conversion to obtain third image data corresponding to a second area on the image sensor and fourth image data for displaying on a display screen, wherein the second area is an area where a specified object is located, the number of pixels of the third image data is M2 × N2, M2 is the number of pixel rows, R is not less than M2 and not more than M, N2 is the number of pixel columns, S is not less than N2 and not more than N, M2 is not equal to M1, N2 is not equal to N1, and the number of pixels of the fourth image data is R × S; and moving the second area in response to the movement of the specified object.
Optionally, the method may further include: in response to a partial or full movement of the second region beyond the range of the first region, expanding and/or moving the range of the first region on the image sensor such that the expanded and/or moved first region includes the second region; and/or in response to partial or whole movement of the second area beyond the range of the first area, rotating the shooting direction of the imaging device to which the image sensor belongs, so that the first area on the rotated image sensor comprises the second area.
According to a seventh aspect of the present disclosure, there is provided an image processing method for processing an image sensed by an image sensor to be displayed on a display screen, wherein the number of pixels of sensor image data sensed by the image sensor is M × N, the number of pixels of image data supplied to the display screen is R × S, R and M are the number of pixel rows, respectively, and R < M, S and N are the number of pixel columns, respectively, and S < N. The method comprises the following steps: performing image conversion to obtain first image data corresponding to a first area on the image sensor and second image data for displaying on a display screen, wherein the pixel number of the first image data is M1 × N1, M1 is the pixel row number, R is not less than M1 and not more than M, N1 is the pixel column number, S is not less than N1 and not more than N, and the pixel number of the second image data is R × S; identifying a specified object from the second image data; performing image conversion to obtain third image data corresponding to a second area on the image sensor and fourth image data used for displaying on the display screen, wherein the second area is an area where a specified object is located, the number of pixels of the third image data is M2 × N2, M2 is the number of pixel rows, R is not less than M2 and not more than M, N2 is the number of pixel columns, S is not less than N2 and not more than N, M2 is not equal to M1, N2 is not equal to N1, and the number of pixels of the fourth image data is R × S; and identifying the identity information of the specified object using a face recognition function based on the fourth image data.
Optionally, the method may further include: based on the identified identity information, granting the authority of the user corresponding to the specified object to execute the predetermined operation; and/or performing security check on the user corresponding to the specified object based on the identified identity information.
According to an eighth aspect of the present disclosure, there is provided an image processing apparatus for processing an image sensed by an image sensor to be displayed on a display screen, wherein the number of pixels of sensor image data sensed by the image sensor is mxn, the number of pixels of image data supplied to the display screen or a device to which the display screen belongs is rxs, R and M are the number of pixel rows, respectively, and R < M, S and N are the number of pixel columns, respectively, and S < N, the method comprising: first reading means for reading first image data corresponding to the first region from the image sensor, the number of pixels of the first image data being M1 × N1, M1 being the number of pixel rows, and R ≦ M1 ≦ M, N1 being the number of pixel columns, and S ≦ N1 ≦ N; a first conversion means for converting the first image data into second image data having a pixel number of R × S; the first sending device is used for sending the second image data to the display screen or the equipment to which the display screen belongs so as to display the image; second reading means for reading third image data corresponding to the second area from the image sensor in response to a zoom instruction, the third image data having a pixel count of M2 × N2, M2 being a pixel row number, and R ≦ M2 ≦ M, N2 being a pixel column number, and S ≦ N2 ≦ N, M2 ≠ M1, N2 ≠ N1; second conversion means for converting the third image data into fourth image data having a pixel count of R × S; and the second sending device is used for sending the fourth image data to the display screen or the equipment to which the display screen belongs so as to display the image.
According to a ninth aspect of the present disclosure, there is provided an image processing apparatus for processing an image sensed by an image sensor to be displayed on a display screen, wherein the number of pixels of sensor image data sensed by the image sensor is mxn, the number of pixels of image data transmitted to the display screen or a device to which the display screen belongs is rxs, R and M are the number of pixel rows, respectively, and R < M, S and N are the number of pixel columns, respectively, and S < N, the method comprising: reading means for reading sensor image data from the image sensor and storing the sensor image data in the memory; first extracting means for extracting first image data corresponding to the first area from the sensor image data on the memory, the number of pixels of the first image data being M1 × N1, M1 being the number of pixel rows, and R ≦ M1 ≦ M, N1 being the number of pixel columns, and S ≦ N1 ≦ N; a first conversion means for converting the first image data into second image data having a pixel number of R × S; the first sending device is used for sending the second image data to the display screen or the equipment to which the display screen belongs so as to display the image; second extracting means for extracting, in response to a zoom instruction, third image data corresponding to the second region from the sensor image data on the memory, the third image data having a pixel number of M2 × N2, M2 being a pixel row number, R ≦ M2 ≦ M, N2 being a pixel column number, and S ≦ N2 ≦ N, M2 ≠ M1, N2 ≠ N1; second conversion means for converting the third image data into fourth image data having a pixel count of R × S; and the second sending device is used for sending the fourth image data to the display screen or the equipment to which the display screen belongs so as to display the image.
According to a tenth aspect of the present disclosure, there is provided an image processing apparatus comprising: the device comprises a camera device, a data acquisition device and a data processing device, wherein the camera device is used for sensing a field image on an image sensor in real time, the pixel number of sensor image data sensed by the image sensor is M multiplied by N, M is the pixel row number, and N is the pixel column number; a display screen for displaying a live image sensed on the image sensor, wherein the number of pixels of image data provided to the display screen is R × S, R is the number of pixel rows, R < M, S is the number of pixel columns, and S < N; and a processor for processing the image data from the image sensor for display on the display screen, wherein the processor reads first image data corresponding to the first region from the image sensor, converts the first image data into second image data, and sends the second image data to the display screen for image display, wherein the first image data has a pixel number of M1 × N1, the second image data has a pixel number of R × S, M1 is a pixel row number, and R ≦ M1 ≦ M, N1 is a pixel column number, and S ≦ N1 ≦ N; in response to a zoom instruction, the processor reads third image data corresponding to the second area from the image sensor, converts the third image data into fourth image data, and sends the fourth image data to the display screen for image display, wherein the number of pixels of the third image data is M2 × N2, M2 is the number of pixel rows, R ≦ M2 ≦ M, N2 is the number of pixel columns, S ≦ N2 ≦ N, M2 ≠ M1, N2 ≠ N1, and the number of pixels of the fourth image data is R × S.
According to an eleventh aspect of the present disclosure, there is provided an image processing apparatus comprising: the device comprises a camera device, a data acquisition device and a data processing device, wherein the camera device is used for sensing a field image on an image sensor in real time, the pixel number of sensor image data sensed by the image sensor is M multiplied by N, M is the pixel row number, and N is the pixel column number; a display screen for displaying a live image sensed on the image sensor, wherein the number of pixels of image data provided to the display screen is R × S, R is the number of pixel rows, R < M, S is the number of pixel columns, and S < N; a memory for storing sensor image data read from the image sensor; and a processor for processing the sensor image data to be displayed on the display screen, wherein the processor extracts first image data corresponding to the first region from the sensor image data on the memory, converts the first image data into second image data, and transmits the second image data to the display screen for image display, wherein the first image data has a pixel number of M1 × N1, the second image data has a pixel number of R × S, M1 is a pixel row number, and R ≦ M1 ≦ M, N1 is a pixel column number, and S ≦ N1 ≦ N; in response to a zoom instruction, the processor extracts third image data corresponding to the second region from the sensor image data on the memory, converts the third image data into fourth image data, and sends the fourth image data to the display screen for image display, wherein the number of pixels of the third image data is M2 × N2, M2 is the number of pixel rows, and R ≦ M2 ≦ M, N2 is the number of pixel columns, and S ≦ N2 ≦ N, M2 ≠ M1, N2 ≠ N1, and the number of pixels of the fourth image data is R × S.
According to a twelfth aspect of the present disclosure, there is provided an image processing apparatus comprising: the device comprises a camera device, a data acquisition device and a data processing device, wherein the camera device is used for sensing a field image on an image sensor in real time, the pixel number of sensor image data sensed by the image sensor is M multiplied by N, M is the pixel row number, and N is the pixel column number; a processor for processing image data from the image sensor for display on a display screen; and a network module for connecting to a network, wherein the processor reads first image data corresponding to the first area from the image sensor, converts the first image data into second image data, and transmits the second image data to an external device having a display screen via the network module for image display, wherein the first image data has a pixel count of M1 × N1, the second image data has a pixel count of R × S, M1 and R are pixel row numbers, respectively, and R ≦ M1 ≦ M, N1 and S are pixel column numbers, respectively, and S ≦ N1 ≦ N; in response to the zoom instruction, the processor reads third image data corresponding to the second region from the image sensor, converts the third image data into fourth image data, and transmits the fourth image data to an external device having a display screen via the network module for image display, wherein the number of pixels of the third image data is M2 × N2, M2 is the number of pixel rows, and R ≦ M2 ≦ M, N2 is the number of pixel columns, and S ≦ N2 ≦ N, M2 ≠ M1, N2 ≠ N1, and the number of pixels of the fourth image data is R × S.
According to a thirteenth aspect of the present disclosure, there is provided an image processing apparatus including: the device comprises a camera device, a data acquisition device and a data processing device, wherein the camera device is used for sensing a field image on an image sensor in real time, the pixel number of sensor image data sensed by the image sensor is M multiplied by N, M is the pixel row number, and N is the pixel column number; a memory for storing sensor image data read from the image sensor; a processor for processing the sensor image data for display on a display screen; and a network module for connecting to a network, wherein the processor extracts first image data corresponding to the first area from the sensor image data on the memory, converts the first image data into second image data, and transmits the second image data to an external device having a display screen via the network module for image display, wherein the first image data has a pixel count of M1 × N1, the second image data has a pixel count of R × S, M1 and R are pixel line numbers, respectively, and R ≦ M1 ≦ M, N1 and S are pixel column numbers, respectively, and S ≦ N1 ≦ N; in response to a zoom instruction, the processor extracts third image data corresponding to the second area from the sensor image data on the memory, converts the third image data into fourth image data, and transmits the fourth image data to an external device having a display screen via the network module for image display, wherein the third image data has a pixel count of M2 × N2, M2 is a pixel row count, and R ≦ M2 ≦ M, N2 is a pixel column count, and S ≦ N2 ≦ N, M2 ≠ M1, N2 ≠ N1, and the fourth image data has a pixel count of R × S.
According to a fourteenth aspect of the present disclosure, there is provided an image processing system comprising: the image processing apparatus according to the seventh or eighth aspect described above; and an electronic device having a display screen, receiving the second image data and the fourth image data from the image processing device, and performing image display accordingly.
Optionally, the image processing apparatus may further include a local display screen, and the image processing apparatus receives an image from the electronic apparatus via the network module and displays the received image on the local display screen.
According to a fifteenth aspect of the present disclosure, there is provided a computing device comprising: a processor; and a memory having executable code stored thereon, which when executed by the processor, causes the processor to perform the method as described in the first aspect above.
According to a sixteenth aspect of the present disclosure, there is provided a non-transitory machine-readable storage medium having stored thereon executable code which, when executed by a processor of an electronic device, causes the processor to perform the method as described in the first aspect above.
Therefore, by intercepting the pixel data corresponding to different area ranges on the image sensor, a clear zooming display effect on the display screen is conveniently realized without additionally increasing the hardware cost.
Drawings
The above and other objects, features and advantages of the present disclosure will become more apparent by describing in greater detail exemplary embodiments thereof with reference to the attached drawings, in which like reference numerals generally represent like parts throughout.
Fig. 1 schematically shows an image processing apparatus of the present disclosure.
Fig. 2 schematically shows an image data processing procedure in the sensor zoom scheme of the present disclosure.
Fig. 3 is a schematic flow diagram of an image processing method according to one embodiment of the present disclosure.
Fig. 4 is a schematic block diagram of an image processing apparatus according to one embodiment of the present disclosure.
Fig. 5 is a schematic flow chart diagram of an image processing method according to another embodiment of the present disclosure.
Fig. 6 is a schematic block diagram of an image processing apparatus according to another embodiment of the present disclosure.
Fig. 7 shows a schematic structural diagram of a computing device which can be used for implementing the image processing method according to an embodiment of the invention.
Detailed Description
Preferred embodiments of the present disclosure will be described in more detail below with reference to the accompanying drawings. While the preferred embodiments of the present disclosure are shown in the drawings, it should be understood that the present disclosure may be embodied in various forms and should not be limited to the embodiments set forth herein. Rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the disclosure to those skilled in the art.
Currently, many apparatuses have a camera or the like capable of taking live images (including still images and/or moving videos) in real time. The images shot by the camera device can be displayed by a display screen of the equipment or a display screen of other equipment after being processed.
In the image processing scheme of the present disclosure, cropping of the angle of view range is achieved by means of sensor crop (sensor crop) or image signal processing crop (ISP crop), thereby achieving a zoom (zoom) function. This type of zooming may be referred to as "sensor zooming" as distinguished from conventional "optical zooming" and "digital zooming".
Fig. 1 schematically shows an image processing apparatus of the present disclosure. In some cases, it may also be referred to as an "image processing system".
As shown in fig. 1, the image processing apparatus/system may include a camera 2, a processor 4, and a display screen 7. In some embodiments, the image processing apparatus/system may further include a memory 6 and a network module 5.
The imaging device 2 may be an imaging apparatus such as a camera, for example, and may capture live images in real time. The image may be a still picture or a moving video including one frame of image.
A person or various possible photographic objects or objects of interest, such as the user 1, may be in the photographic field and photographed by the camera 2.
The imaging device 2 adopted by the present disclosure may employ, for example, a wide-angle lens, and can clearly image an object in a field space within a wide angle of view.
As an example, fig. 1 schematically shows a shooting angle range limit 20 of the imaging device 2 in a dashed line. The size of the imaging angle of view range may be, for example, the vertex angle of a cone surrounded by the limit. For example, the shooting angle of view range of the image pickup device 2 may be 120 °, or a larger shooting angle of view range may be also possible if the apparatus permits.
The imaging device 2 has an image sensor 3, and a live image is sensed on the image sensor 3, for example, in real time. For simplicity of illustration and the following description, it may be considered that the photosensitive area of the image sensor 3 is shown in fig. 1.
In fig. 1, (the photosensitive area of) the image sensor 3 is shown enlarged in the lower left corner, as well as an image 10 of the object 1 formed on the image sensor 3 (e.g. in the form of data corresponding to individual sensor pixels).
In fig. 1, an imaging area limit 200 of the photographing view angle range limit 20 on the plane of the image sensor 3 is shown in the lower left corner by a dotted circle bordering (the photosensitive area of) the image sensor 3, corresponding to the photographing view angle range limit 20 shown in the upper left corner by a dotted ray.
It should be understood that the dashed circle of the imaging area boundary 200 may also not border (the photosensitive area of) the image sensor 3. Alternatively, the imaging area may be larger and fully encompass the image sensor 3. In this way, (the photosensitive area of) the image sensor 3 can sense and obtain the sensor image data 30 obtained by sensing and shooting the scene.
Alternatively, in some cases, the imaging area may be smaller, the imaging area boundary 200 intersects with the edge of (the photosensitive area of) the image sensor 3, and the four corner portions of the image sensor 3 will not fall within the imaging area, cannot receive the field light collected by the camera device 2, and cannot be used to form the sensor image data 30 of the field. In other words, the sensed sensor image data 30 may correspond to a size that is less than the full photosensitive area size of the image sensor 3.
The resolution of the image sensor 3 may be selected to match the resolution of the lens.
The resolution of the lens is defined as the number of pairs of black and white stripes that the lens can resolve within a unit millimeter at the image plane.
The resolution of the image sensor 3 depends on its photosensitive area size (chip size) and single pixel size. The resolution of the image sensor can be expressed in terms of the number of effective pixels of the image sensor.
With the development of lens technology and image sensor technology, the effective pixel number of the image sensor can reach millions of pixels, for example, 500 pixels, that is, the pixel number of each row and each column can be about 2000 pixels, for example, 2592 × 1944 or 2560 × 1920 can be achieved.
Generally, for convenience of description, the number of pixels of the sensor image data 30 sensed by the image sensor 3 may be M × N, M being the number of pixel rows and N being the number of pixel columns.
The processor 4 processes the image data from the image sensor 3 for display on the display screen 7.
The processor 4 may read all or part of the image data directly from the image sensor 3 for processing for presentation to the display screen 7 in accordance with aspects of the present disclosure as will be described below. In this case, all or part of the image data read from the image sensor 3 may also be stored in the memory 6 for processing.
Alternatively, the processor 4 may first read all the image data on the image sensor 3 and then write the image data into the memory 6, and then read all or part of the image data from the memory 6 for processing according to the scheme of the present disclosure to be described later, so as to send the image data to the display 7 for display.
The display screen 7 displays a live image 300 sensed on the image sensor 3 based on the image data sent from the processor 4. The display image 100 of the object 1 may be included in the live image 300.
Although many large-sized high-definition display screens, such as 4K display screens, have been available at present, the resolution can reach 3840 × 2160 or 4096 × 2160, the display screens of the common devices do not have such high resolution. Especially for presenting live images, tend to have smaller sizes and lower resolutions. For example, many scenarios employ display screens having resolutions of only 640 × 480 or 800 × 600. Also, many times, not all of the display area on the display screen 7 is used for displaying the live image 300, but only a portion of the display area on the display screen 7 displays the live image 300.
Thus, the number of pixels required for the display screen 7 to display the live image 300 will often be significantly less than the number of pixels sensed on the image sensor 3.
The number of pixels of the image data that needs to be provided to the display screen 7 or the device to which the display screen 7 belongs may be equal to or greater than the number of pixels of the portion of the display area of the display screen 7 that is used to display the live image 300.
In the case where the number of pixels of the image data provided to the display screen 7 is greater than the number of pixels of the portion of the display area of the display screen 7 for displaying the live image 300, the image processing means of the display screen 7 itself or the image processing means associated with the display screen 7 may perform scaling processing on the image data to obtain image data having the number of pixels equal to the number of pixels of the display area, so as to be displayed by the portion of the display area of the display screen 7.
If display is to be performed on a display area of the display screen 7 having a large number of pixels based on image data having a small number of pixels, the number of pixels needs to be increased by processing such as image interpolation. This may reduce the sharpness of the displayed image, may result in some blurred contours, etc.
On the other hand, if the image data having a large number of pixels is displayed in the display area of the display panel 7 having a small number of pixels, the display capability of the display panel can be fully utilized, and the resolution of the displayed image is not reduced within the range allowed by the resolution of the display panel.
With the image sensor 3 having a higher resolution (larger number of pixels) and the display screen 7 having a lower resolution, an operation space is provided for image processing.
Generally, for convenience of description, the number of pixels of image data to be supplied to the display screen 7 or a device to which the display screen 7 belongs may be represented as R × S, R being the number of pixel rows and S being the number of pixel columns.
As described above, the number R of pixel rows of the image data to be supplied to the display screen 7 is smaller than the number M of pixel rows of the image sensor 3, and the number S of pixel columns of the image data to be supplied to the display screen 7 is smaller than the number N of pixel columns of the image sensor 3.
The camera 2 (and its image sensor 3), the processor 4, the memory 6 and the display 7 may be integrated on one and the same device. Alternatively, the image processing/display system may be provided as separate devices or may be provided on different devices, thereby forming a real-time live image processing/display system.
The display 7 may be provided near the site (which may be referred to as "local") where the camera 2 is taking pictures, for example for viewing by the user 1 or others at the site.
Alternatively, the display 7 may be located far from the shooting site of the imaging apparatus 2, and the display 7 may be said to be in a "different place", that is, a "different place display".
At this time, the image pickup device 2 (and the image sensor 4 thereof), the processor 4, the network module 5, and the memory 6 may be provided as separate devices, respectively, or any two or more thereof may be included in the same device, and may be referred to as an "image processing device", for example. The combination of the image processing device and the displaced display 7 or the device to which the displaced display 7 belongs may be referred to as an "image processing system".
Also in this case, the display 7 may be a separate device or may belong to another device, and may generally be an electronic device, such as a smart mobile phone or the like.
The image data is transmitted via the network module 5, for example a router, over the network to the display 7 or to the device to which the display 7 belongs.
In addition, the camera associated with the displaced display 7 or the camera of the device to which the displaced display 7 belongs can also take images of its vicinity, which can for example comprise images of the respective user, and transmit them via the network module 5, which takes images of the vicinity of the site, via the network back to the local (local image processing device), for example received by the processor 4 and, if necessary, stored on the memory 6.
A local display (not shown in fig. 1, unlike the display 7 described above) may also be provided locally. The image taken by the camera associated with the displaced display screen 7 or the camera of the apparatus to which the displaced display screen 7 belongs can be displayed on the local display screen.
In this way, a displaced user can view an image of the local scene on his display 7. Alternatively, a displaced user may have a local user engaged in a video call.
The various components shown in fig. 1 may be combined in various ways.
For example, the camera 2 (and its image sensor 3), the display 7 and the processor 4 may be integrated in one image processing apparatus. The processor 4 reads image data from the image sensor 3 and performs image processing according to the present disclosure, and displays the processed image on the display screen 7.
Alternatively, the camera 2 (and its image sensor 3), the display 7, the memory 6 and the processor 4 may be integrated in one image processing apparatus. The image data read from the image sensor 3 is stored in the memory 6. The processor 4 extracts the corresponding image data from the memory 6 and performs image processing according to the present disclosure, and displays the processed image on the display 7.
Alternatively, the camera 2 (and its image sensor 3), the processor 4 and the network module 5 may be integrated in one image processing apparatus. The processor 4 reads image data from the image sensor 3 and performs image processing according to the present disclosure. The processed image data may be transmitted to an external device having a display screen via the network module 5 for image display. The image processing apparatus and the external apparatus may be collectively referred to as an image processing system.
Alternatively, the camera 2 (and its image sensor 3), the memory 6, the processor 4 and the network module 5 may be integrated in one image processing apparatus. The image data read from the image sensor 3 is stored in the memory 6. The processor 4 extracts the respective image data from the memory 6 and performs image processing according to the present disclosure. The processed image data may be transmitted to an external device having a display screen via the network module 5 for image display. The image processing apparatus and the external apparatus may be collectively referred to as an image processing system.
An image processing method for realizing the above-described "sensor zoom" according to the present disclosure is described below with reference to fig. 2 to 6. The image processing method may be performed by the processor 4 described above.
Fig. 2 schematically shows an image data processing procedure in the sensor zoom scheme of the present disclosure.
An image 315 (to the right) including the displayed image 110 of the object 1 may be displayed based on pixel data in a first region 310 on the image sensor 3 corresponding to the viewing angle range 210.
Here, the first region 310 may be the entire effective region of the image sensor 3 or a part thereof. When the first area 310 is the entire effective area or a large portion of the area of the image sensor 3, the displayed image 110 of the object 1 may be smaller in the entire displayed image 315, and it is inconvenient to view the details of the object 1.
In response to the zoom instruction, the (lower side) image 325 including the display image 120 of the object 1 may be displayed based on the pixel data in the second area 320 corresponding to the angle of view range 220 on the image sensor 3 again.
Where the zoom instruction is intended to zoom in on the displayed image of object 1, the second region 320 may be selected to be smaller than the first region 310. Alternatively, the viewing angle range 220 is less than the viewing angle range 210. For example, where the viewing angle range 210 corresponds to a 120 ° viewing angle range, the range of the viewing angle range 220 may correspond to a 60 ° viewing angle range.
In this way, the display image 120 of the object 1 is increased in proportion to the entire displayed image 325. When image 325 is displayed in the same size as image 315, displayed image 120 of object 1 may be enlarged as compared to displayed image 110, achieving a zoom effect.
The magnification of the zoom may be determined in accordance with the zoom instruction, or may be set to a fixed zoom magnification, for example, the angle of view range is reduced to half of the original, or the size of the display content is enlarged to twice the original.
The number of pixels M x N of the image sensor 3 may be significantly larger than the number of pixels R x S required by the display screen 7 to display live images (or the number of pixels that need to be sent to the display screen 7 or to the device to which it belongs). In this way, it is possible to make the number of pixels larger than the number of pixels R × S required for the display screen 7 to realize clear display, in either the first region 310 or the second region 320.
An image processing method according to an embodiment of the present disclosure is described below with reference to fig. 3 and 4.
Fig. 3 is a schematic flow diagram of an image processing method according to one embodiment of the present disclosure.
Fig. 4 is a schematic block diagram of an image processing apparatus according to one embodiment of the present disclosure.
As shown in fig. 4, the image processing apparatus 400 may include a first reading device 410, a first conversion device 420, a first transmitting device 430, a second reading device 440, a second conversion device 450, and a second transmitting device 460.
In step S310, first image data corresponding to the first region 310 may be read from the image sensor 3, for example, by the first reading device 410.
It should be understood that in the case of displaying a still picture, the first image data on the image sensor 3 may not change over time.
In the case of displaying a dynamic video, the first image data on the image sensor 3 may be changed at any time according to a change in the field situation. In the case of displaying a dynamic video, the image sensor 3 senses an image of one frame, and the processor 4 acquires image data of one frame from the image sensor 3 and then displays the image of one frame on the display screen 7, thereby displaying the dynamic video. Only as described below, when zoom display is required, the area range in which pixel data is read from the image sensor 3 can be changed.
As described above, the first region 310 may be the image sensor 3 or all of the effective pixel region thereof, or may be a partial region on the image sensor 3.
If initially displayed, it may be selected to include as large an area as possible. For example, the first region 310 may include the entire image sensor 3 or all of its effective pixel region.
The number of pixels of the first image data (corresponding to the first region 310) can be recorded as M1 Xn 1, M1 is the number of pixel rows, R ≦ M1 ≦ M, N1 is the number of pixel columns, and S ≦ N1 ≦ N. In other words, the number of pixel rows M1 and the number of pixel columns N1 of the first image data (corresponding to the first area 310) are respectively between the number of pixel rows M and the number of pixel columns N of the image sensor 3 and the number of pixel rows R and the number of pixel columns S of the image data that needs to be provided to the display screen 7 or the device thereof. In the case of M1< M, N1< N, such an operation may be referred to as "sensor cut (sensorrop)".
In this way, sufficient pixel data is provided for the following image processing, so that image display can be performed with the display capability of the display screen 7 being fully utilized.
In step S320, the first image data (corresponding to the first area 310) may be converted into the second image data 315, for example, by the first conversion means 420, and the number of pixels of the second image data 315 is the above-mentioned number R × S of pixels to be provided to the display screen 7 or the device thereof.
The image scaling conversion process is well known in the art and will not be described herein.
Since the number of pixels m1 × n1 of the first image data (310) is not smaller than the number of pixels of the second image data 315, which is R × S in the technical aspect of the present disclosure, and there is sufficient pixel data with respect to the display capability of the display screen 7, only the "reduction" processing may be performed here without performing the "enlargement" processing such as interpolation. After the image conversion in step S320, the display capability of the display screen 7 can still be fully utilized, and the definition of the displayed image 315 is not reduced within the range allowed by the resolution of the display screen 7.
Then, in step S330, the second image data 315 may be transmitted to the display screen 7 or the device to which the display screen 7 belongs for image display, for example, by the first transmitting means 430. Accordingly, in the displayed image, the image 110 of the object 1 may be displayed.
It should be understood that in the case where the first image data is moving video data, the conversion processing of step S320 and the transmission of step S330 are performed for the first image data of one frame by one frame. In other words, here, transmitting the second image data 315 refers to continuously transmitting the dynamically changing second image data obtained based on the dynamically changing first image data.
In displaying the image 110, a zoom instruction may be issued.
The zoom instruction may be issued in response to a user operation, such as a zoom operation by the user on a touch display screen, or a drag operation by the user on a zoom ratio control bar.
In this way, zoom instructions may be received from the device and/or server to which the display belongs.
Alternatively, the zoom instruction may be automatically generated. For example, a zoom instruction may be triggered in response to identifying a specified object. For example, the specified object may be an image sensed by the image sensor 3 for a specified object, such as a human face image.
Specifically, for example, a specified object (e.g., the image 110 of the object 1) may be identified from the second image data 315. In response to identifying the specified object in the second image data 315, a zoom instruction is issued.
When the specified object is a face image, the specified object may be identified from the second image data using a face recognition function.
In consideration of the requirement of privacy protection of the user, a function switch can be set, and the face recognition function is turned on or off in response to the setting of the user.
When the face image occupies a relatively small area in the second image data 315, the face image can be generally recognized, although the identity information corresponding to the face image is not necessarily obtained.
The zoom instruction may be issued in a case where a distance from a specified object, such as a face image, to the imaging device to which the image sensor belongs is greater than or equal to a first distance threshold.
On the other hand, in the case where the distance from the specified object, such as a face image, to the imaging apparatus to which the image sensor belongs is less than the first distance threshold, then the zoom instruction may not be issued.
The first distance threshold may be set based on experience and/or parameters of the imaging apparatus (image pickup device). For example, the first distance threshold may be set such that when the distance is greater than or equal to the first distance threshold, the image of the designated object will become less visible, and a zoom operation is required to enlarge the image of the designated object.
For example, in the case of a home camera, the first distance threshold may be, for example, 3 meters or 5 meters.
The magnification of zooming may be a preset value or may be automatically determined according to, for example, the size ratio occupied by the designated image in the second image data 315.
In other words, the number of pixels m2 × n2 of the third image data, that is, the zoom magnification may be determined based on the screen aspect ratio of the designated object with respect to the first region and the number of pixels m1 × n1 of the first image data.
In addition, the zoom instruction may be issued in a case where the screen occupation ratio of the specified object such as a face image with respect to the first area is less than or equal to a first occupation ratio threshold.
On the other hand, in the case where the screen aspect ratio of the specified object such as a face image with respect to the first area is larger than the first aspect ratio threshold, the zoom instruction may not be issued.
The first screen ratio may be set based on experience and/or parameters of the display screen, for example. For example, the first screen duty may be set such that when the screen duty of the designated object with respect to the first area is less than or equal to the first duty threshold, the image of the designated object is small and not easy to observe, and then a zoom operation is required to enlarge the image of the designated object.
In step S340, in response to the zoom instruction, the third image data corresponding to the second region 320 may be read from the image sensor 3 by, for example, the second reading device 440.
The center of the second region 320 may coincide with the center of the first region 310.
Alternatively, the center position of the second region 320 may also vary with respect to the center position of the first region 310.
For example, the center position of the second region 320 may be determined according to a zoom instruction.
For example, when a zoom instruction is issued in response to a finger zoom operation on the touch display screen, the center position of the second area 320 is determined based on the center positions of the plurality of finger operations. Alternatively, the center position of the second area 320 is determined by receiving a click operation of the user.
The second area 320 may be a designated area in the first area 310. Alternatively, in some cases, the second region 320 may also be a region overlapping with the first region 310. Alternatively, the second area 320 page may contain a larger area of the first area 310 (e.g., when the first area 310 is already a smaller range on the image sensor 3 cropped according to the sensor zoom scheme of the present disclosure).
Alternatively, for example, in the case where a zoom instruction is triggered in response to recognition of the designated object as described above, the area in which the designated object is located may be set as the second area 320.
In addition, the second region 320 may also be moved accordingly as the designated object moves.
When part or all of the second region is moved beyond the range of the first region, the range of the first region on the image sensor may be expanded and/or moved if the range of the first region is less than the total pixel range of the image sensor, such that the expanded and/or moved first region includes the second region.
Alternatively, when part or all of the second area moves beyond the range of the first area, the shooting direction of the imaging device to which the image sensor belongs may also be rotated toward the direction in which the second area moves, so that the first area on the image sensor includes the second area after the rotation.
The number of pixels of the third image data can be recorded as M2 × N2, M2 is the number of pixel rows, R is equal to or less than M2 and equal to or less than M, N2 is the number of pixel columns, S is equal to or less than N2 and equal to or less than N, M2 is equal to M1, and N2 is equal to N1. In other words, in response to a zoom instruction, the area range of pixel data that needs to be read from the image sensor 3 is different from before.
When the zoom instruction indicates that it is desired to enlarge the displayed object, the number of pixels of the third image data is smaller than that of the first image data, i.e., m2< m1, n2< n 1.
When the zoom instruction indicates that it is desired to reduce the displayed object, the number of pixels of the third image data is larger than that of the first image data, i.e., m2> m1, n2> n 1.
As with the first image data, the third image data (corresponding to the second region 320) has a number M2 of pixel rows and a number N2 of pixel columns between the number M of pixel rows and the number N of pixel columns of the image sensor 3 and the number R of pixel rows and the number S of pixel columns of the image data that needs to be supplied to the display screen 7 or the device thereof, respectively.
Like the above step S320, in step S350, the third image data (corresponding to the second area 320) may be converted into the fourth image data 325 by the second conversion means 450, for example, and the number of pixels of the fourth image data 325 is the above-mentioned number R × S of pixels to be supplied to the display 7 or the device thereof.
Then, in step S360, the fourth image data 325 may be transmitted to the display screen 7 or the device to which the display screen 7 belongs for image display, for example, by the second transmitting means 460. Accordingly, in the displayed image, the image 120 of the object 1 may be displayed.
In this way, since the range of the second region 320 corresponding to the fourth image data 325 supplied to the display 7 on the image sensor 3 is changed from the range of the previous first region 310, the ratio of the image 10 of the object 1 on the image sensor 3 in the cut-out region (first region 310/second region 320) is changed accordingly.
For example, in the case where the zoom operation is performed to enlarge the image display of the object 1, when the second image data 315 corresponding to the first area 310 is previously displayed on the display screen 7 in a certain display area (the entire display screen or a part thereof), the display image 110 of the object 1 is small, and after the above-described sensor zoom operation is performed, when the fourth image data 325 corresponding to the second area 320 is displayed on the display screen 7 in the same display area, the display image 115 of the object 1 may be relatively enlarged.
It should be understood that in the case where the third image data is moving video data, the conversion processing of step S350 and the transmission of step S360 are performed for the third image data of one frame. In other words, here, transmitting the fourth image data 325 means continuously transmitting the dynamically changing fourth image data obtained based on the dynamically changing third image data.
For example, corresponding parameters may be determined by the above zoom operation, and then pixels on the image sensor may be processed with the same parameters for each subsequent frame of image. That is, the third image data in the same range is read for each frame of the original image data on the image sensor, and the fourth image data corresponding to the same number of pixels in the second area is obtained by performing the scaling processing in the same scale.
Therefore, when zooming processing is required, clear zooming display effect on the display screen 7 is realized by intercepting pixel data in different area ranges on the image sensor 3.
As described above, in the second image data, it is possible that since the picture proportion of a specified object such as a face image is small, although a face can be recognized, it is not sufficient to perform identity analysis by face recognition.
After passing through the above-described zoom processing, the screen aspect ratio of a specified object such as a face image in the fourth image data increases, and the face image is enlarged. In the above-mentioned enlarging process according to the present disclosure, the amount of face image information (image details), i.e., the number of pixels of the face image portion, is also increased at the same time.
Then, the identity information of the specified object may be recognized using the face recognition function based on the fourth image data.
In this way, the identified identity information may be used in various application scenarios.
For example, a right to perform a predetermined operation may be granted to a user corresponding to a specified object, such as a face image, based on the recognized identity information. E.g. logging on to a certain system, obtaining a certain information or content, entering a certain physical space, etc.
For another example, a security check may be performed on a user corresponding to a designated object, such as a face image, based on the identified identity information. For example, identity comparisons and the like can be performed.
Thereafter, if the zoom processing needs to be performed again, the operations of step S340 to step S360 described above may be performed again.
In addition, after the fourth image data is transmitted, in other words, the image processing is performed using the zoomed parameter, and after the image data is transmitted and displayed, in response to a distance from the specified object to the imaging device to which the image sensor belongs becoming smaller than the second distance threshold, the second image data may also be transmitted again to the display screen or the device to which the display screen belongs for image display. That is, the display state before zooming is restored, image processing is performed using the parameter before zooming, and image data is transmitted and displayed.
Alternatively, after the fourth image data is transmitted, in other words, the image processing is performed using the zoomed parameter, and after the image data is transmitted and displayed, in response to the screen aspect ratio of the specified object with respect to the first area becoming larger than the second aspect ratio threshold, the second image data may also be transmitted again to the display screen or the device to which the display screen belongs for image display. That is, the display state before zooming is restored, image processing is performed using the parameter before zooming, and image data is transmitted and displayed.
The above describes a technical solution of directly reading the pixel data of the trimming rear partial area from the image sensor 3 when the trimming operation is performed.
In another embodiment, image data of all or all of the active pixels may also be read from the image sensor 3 and stored in the memory 6. Then, in the following process, the image data corresponding to the respective areas, i.e., the pixel data in the respective ranges, is extracted from the memory 6.
An image processing method according to another embodiment of the present disclosure is described below with reference to fig. 5 and 6.
Fig. 5 is a schematic flow chart diagram of an image processing method according to another embodiment of the present disclosure.
Fig. 6 is a schematic block diagram of an image processing apparatus according to another embodiment of the present disclosure.
As shown in fig. 6, the image processing apparatus 600 may include a first reading device 610, a first converting device 620, a first transmitting device 630, a second reading device 640, a second converting device 650, and a second transmitting device 660.
The image processing method shown in fig. 5 has one more step S505 of reading and storing data than the image processing method shown in fig. 3. Accordingly, in steps S510 and S540, image data corresponding to the respective areas are extracted from the memory. The other steps S520, S530, S550, S560 may be substantially the same as S320, S330, S350, S360 described above with reference to fig. 3. Some similarities are not described in detail here.
In step S505, the sensor image data may be read from the image sensor 3, for example, by the reading device 605, and stored in the memory 6. Here, image data sensed by the entire image sensor 3, or image data including pixel data of all effective pixels may be read out from the convex window sensor 3 and stored in the memory 6.
In the case of real-time presentation of a dynamic video, each time a frame of image is formed by the image sensor 3, the frame of image is read and stored in the memory 6, and then the pixel data within the corresponding area is read from the memory 6, processed by the processor 4 and sent to the display screen 7 for display. Thereby enabling the display of real-time dynamic video.
In step S510, first image data corresponding to the first region 310 may be extracted from the sensor image data on the memory 6, for example, by the first extraction means 610, the number of pixels of the first image data may be recorded as M1 × N1, M1 is the number of pixel rows, R ≦ M1 ≦ M, N1 is the number of pixel columns, and S ≦ N1 ≦ N.
In step S520, the first image data may be converted into the second image data 315, for example, by the first conversion means 620, and the number of pixels of the second image data 315 is the above-mentioned number R × S of pixels to be provided to the display 7 or the device thereof.
In step S530, the second image data 315 may be transmitted to the display 7 or the device to which the display 7 belongs for image display, for example, by the first transmitting means 630. Accordingly, in the displayed image, the image 110 of the object 1 may be displayed.
In step S540, in response to the zoom instruction, third image data corresponding to the second region 320 may be extracted from the sensor image data on the memory 6, for example, by the second extraction means 640, the number of pixels of the third image data being M2 × N2, M2 being the number of pixel rows, and R ≦ M2 ≦ M, N2 being the number of pixel columns, and S ≦ N2 ≦ N, M2 ≠ M1, N2 ≠ N1.
In step S550, the third image data is converted into fourth image data 325, and the number of pixels of the fourth image data 325 is R × S.
The fourth image data 325 is sent to the display screen 7 or the device to which the display screen 7 belongs for image display. Accordingly, in the displayed image, the image 120 of the object 1 may be displayed.
Thereby, by changing the range of the image sensor area corresponding to the extracted pixel data when the image data is extracted from the memory 6, a clear zoom display effect on the display screen 7 is achieved.
In addition, for example, in the case where the second area is a part of the first area, in response to a zoom instruction, while the above-described steps S340 to S360 are performed, or in other words, before the fourth image data 325 is transmitted to the display screen 7 or the device to which the display screen 7 belongs in step S360, a dynamic change process from displaying the image corresponding to the first area 310 to displaying the image corresponding to the second area 320 may also be presented on the display screen 7 through a digital zoom process based on the second image data 315 that has been previously received on the display screen 7 side.
During this dynamic change, the area of the image sensor 3 corresponding to the image displayed on the display screen 7 gradually decreases from the larger first area 310 to the smaller second area 320. Each content in the displayed image, for example, the image of the object 1, is gradually enlarged.
Before receiving the fourth image data 325 corresponding to the second region 320, the number of pixels m2 × n2 not less than the number of pixels R × S required for image display of the display screen 7, which is transmitted at step S360, the display screen 7 or the device thereof has the second image data 315 corresponding to the first region 310. The image data corresponding to the second area 320 is also included, except that the number of pixels of the image data is smaller than the number of pixels R × S required for image display by the display panel 7. The displayed image can be enlarged by a digital zooming method through an image data processing method such as interpolation, and the dynamic change process of the zooming enlargement display can be presented.
In addition, in the case that the second area 320 is larger than the first area 310, or a partial area of the second area 320 exceeds the first area 310, the above dynamic change process may also be presented on the display 7. In this way, when the above dynamic change process is displayed, predetermined pixel data, for example, blank pixel data, may be filled in a portion beyond the first region 310, or pixel data calculated based on pixel data of a part or all of the pixels in the first region 310 according to a predetermined rule may be filled in.
Then, in response to the reception of the fourth image data 325 by the display screen 7 or the device to which the display screen 7 belongs at step S360, the above-described dynamic change course display is ended, and image display is performed based on the fourth image data 325.
Thereafter, if the zoom processing needs to be performed again, the operations of step S540 to step S560 described above may be performed again.
The above focuses on the "sensor zoom" scheme according to the present disclosure.
However, one skilled in the art, after reading the above description, may also derive some other embodiments. Some of the details of the content described below may be similar or identical to the details of the corresponding content described above.
For example, in some embodiments, an image processing method may be provided in which, in response to an interactive request, a specified object is identified from image data; and executing zoom processing to increase the screen occupation ratio of the designated object in the image data after the zoom processing in a case where the screen occupation ratio of the designated object in the image data is smaller than a predetermined occupation ratio threshold.
The specified object may be a face image. A face recognition function may be used to identify the specified object from the second image data.
In addition, the face recognition function can be turned on or off in response to the setting of the user.
In addition, in some embodiments, an image processing method for processing an image sensed by an image sensor to be displayed on a display screen may also be provided.
Here, the number of pixels of sensor image data sensed by the image sensor is M × N, the number of pixels of image data supplied to the display screen is R × S, R and M are the number of pixel rows, respectively, and R < M, S and N are the number of pixel columns, respectively, and S < N.
First, image conversion is performed to obtain first image data corresponding to a first area on the image sensor and second image data for display on the display screen. The number of pixels of the first image data is M1 Xn 1, M1 is the number of pixel rows, R is not less than M1 is not less than M, N1 is the number of pixel columns, and S is not less than N1 is not less than N. The number of pixels of the second image data is R × S.
And when the distance from the specified object to the imaging device to which the image sensor belongs is greater than or equal to a first distance threshold value, sending out a zooming instruction.
Then, in response to a zoom instruction, image conversion is performed, resulting in third image data corresponding to a second area on the image sensor and fourth image data for display on the display screen. The second area is an area where the specified object is located. The number of pixels of the third image data is M2 × N2, M2 is the number of pixel rows, R is not less than M2 and not more than M, N2 is the number of pixel columns, S is not less than N2 and not more than N, M2 is not equal to M1, and N2 is not equal to N1. The number of pixels of the fourth image data is R × S.
Further, after the fourth image data is transmitted to the display screen or the device to which the display screen belongs for image display, the second image data may be further transmitted to the display screen or the device to which the display screen belongs again for image display in response to the distance from the specified object to the imaging device to which the image sensor belongs becoming smaller than the second distance threshold.
In addition, in some embodiments, an image processing method for processing an image sensed by an image sensor to be displayed on a display screen may also be provided.
Also, here, the number of pixels of sensor image data sensed by the image sensor is M × N, the number of pixels of image data supplied to the display screen is R × S, R and M are the number of pixel rows, respectively, and R < M, S and N are the number of pixel columns, respectively, and S < N.
First, image conversion is performed to obtain first image data corresponding to a first area on the image sensor and second image data for display on the display screen. The number of pixels of the first image data is M1 Xn 1, M1 is the number of pixel rows, R is not less than M1 is not less than M, N1 is the number of pixel columns, and S is not less than N1 is not less than N. The number of pixels of the second image data is R × S.
When the screen occupation ratio of the designated object with respect to the first area is less than or equal to a first occupation ratio threshold, a zoom instruction is issued.
Then, in response to a zoom instruction, image conversion is performed, resulting in third image data corresponding to a second area on the image sensor and fourth image data for display on the display screen. The second area is an area where the specified object is located. The number of pixels of the third image data is M2 × N2, M2 is the number of pixel rows, R is not less than M2 and not more than M, N2 is the number of pixel columns, S is not less than N2 and not more than N, M2 is not equal to M1, and N2 is not equal to N1. The number of pixels of the fourth image data is R × S.
Further, after the fourth image data is transmitted to the display screen or the device to which the display screen belongs for image display, in response to the screen occupation ratio of the specified object with respect to the first area becoming larger than the second occupation ratio threshold, the second image data is transmitted again to the display screen or the device to which the display screen belongs for image display.
Optionally, the method may further include: the number of pixels m2 × n2 of the third image data is determined based on the screen aspect ratio of the specified object with respect to the first area and the number of pixels m1 × n1 of the first image data.
In addition, in some embodiments, an image processing method for processing an image sensed by an image sensor to be displayed on a display screen may also be provided.
Also, here, the number of pixels of sensor image data sensed by the image sensor is M × N, the number of pixels of image data supplied to the display screen is R × S, R and M are the number of pixel rows, respectively, and R < M, S and N are the number of pixel columns, respectively, and S < N.
First, image conversion is performed to obtain first image data corresponding to a first area on the image sensor and second image data for display on the display screen. The number of pixels of the first image data is M1 Xn 1, M1 is the number of pixel rows, R is not less than M1 is not less than M, N1 is the number of pixel columns, and S is not less than N1 is not less than N. The number of pixels of the second image data is R × S.
Next, the specified object may be identified from the second image data.
Then, in response to a zoom instruction, image conversion is performed, resulting in third image data corresponding to a second area on the image sensor and fourth image data for display on the display screen. The second area is an area where the specified object is located. The number of pixels of the third image data is M2 × N2, M2 is the number of pixel rows, R is not less than M2 and not more than M, N2 is the number of pixel columns, S is not less than N2 and not more than N, M2 is not equal to M1, and N2 is not equal to N1. The number of pixels of the fourth image data is R × S.
It is possible for a given object to move around, for example, in a home or office. The second area may be moved in response to the movement of the specified object, keeping the image of the specified object located in the second area.
As the designated object moves, the second area may also move beyond the range of the first area.
In response to some or all of the second region moving beyond the range of the first region, the range of the first region on the image sensor may be expanded and/or moved such that the expanded and/or moved first region includes the second region.
Alternatively, in response to part or all of the second area moving beyond the range of the first area, the shooting direction of the imaging device to which the image sensor belongs may also be rotated so that the first area on the image sensor after rotation includes the second area.
In addition, in some embodiments, an image processing method for processing an image sensed by an image sensor to be displayed on a display screen may also be provided.
Also, here, the number of pixels of sensor image data sensed by the image sensor is M × N, the number of pixels of image data supplied to the display screen is R × S, R and M are the number of pixel rows, respectively, and R < M, S and N are the number of pixel columns, respectively, and S < N.
First, image conversion is performed to obtain first image data corresponding to a first area on the image sensor and second image data for display on the display screen. The number of pixels of the first image data is M1 Xn 1, M1 is the number of pixel rows, R is not less than M1 is not less than M, N1 is the number of pixel columns, and S is not less than N1 is not less than N. The number of pixels of the second image data is R × S.
Next, the specified object may be identified from the second image data.
Then, image conversion may be performed, resulting in fourth image data corresponding to the third image data of the second area on the image sensor and for display on the display screen. The second area is an area where the specified object is located. The number of pixels of the third image data is M2 × N2, M2 is the number of pixel rows, R is not less than M2 and not more than M, N2 is the number of pixel columns, S is not less than N2 and not more than N, M2 is not equal to M1, and N2 is not equal to N1. The number of pixels of the fourth image data is R × S.
In this way, the identity information of the specified object can be recognized using the face recognition function based on the fourth image data.
Therefore, the authority for executing the predetermined operation of the user corresponding to the specified object can be granted based on the identified identity information.
Or, based on the identified identity information, security check may be performed on the user corresponding to the specified object.
Fig. 7 shows a schematic structural diagram of a computing device which can be used for implementing the image processing method according to an embodiment of the invention.
Referring to fig. 7, computing device 700 includes memory 710 and processor 720.
Processor 720 may be a multi-core processor or may include multiple processors. In some embodiments, processor 720 may include a general-purpose host processor and one or more special purpose coprocessors such as a Graphics Processor (GPU), Digital Signal Processor (DSP), or the like. In some embodiments, processor 720 may be implemented using custom circuits, such as an Application Specific Integrated Circuit (ASIC) or a Field Programmable Gate Array (FPGA).
The memory 710 may include various types of storage units, such as system memory, Read Only Memory (ROM), and permanent storage. Wherein the ROM may store static data or instructions that are required by processor 720 or other modules of the computer. The persistent storage device may be a read-write storage device. The persistent storage may be a non-volatile storage device that does not lose stored instructions and data even after the computer is powered off. In some embodiments, the persistent storage device employs a mass storage device (e.g., magnetic or optical disk, flash memory) as the persistent storage device. In other embodiments, the permanent storage may be a removable storage device (e.g., floppy disk, optical drive). The system memory may be a read-write memory device or a volatile read-write memory device, such as a dynamic random access memory. The system memory may store instructions and data that some or all of the processors require at runtime. In addition, the memory 710 may include any combination of computer-readable storage media, including various types of semiconductor memory chips (DRAM, SRAM, SDRAM, flash memory, programmable read-only memory), magnetic and/or optical disks, may also be employed. In some embodiments, memory 710 may include a removable storage device that is readable and/or writable, such as a Compact Disc (CD), a digital versatile disc read only (e.g., DVD-ROM, dual layer DVD-ROM), a Blu-ray disc read only, an ultra-dense disc, a flash memory card (e.g., SD card, min SD card, Micro-SD card, etc.), a magnetic floppy disk, or the like. Computer-readable storage media do not contain carrier waves or transitory electronic signals transmitted by wireless or wired means.
The memory 710 has stored thereon executable code that, when processed by the processor 720, causes the processor 720 to perform the image processing methods described above.
The image processing scheme according to the present invention has been described in detail above with reference to the accompanying drawings.
Furthermore, the method according to the invention may also be implemented as a computer program or computer program product comprising computer program code instructions for carrying out the above-mentioned steps defined in the above-mentioned method of the invention.
Alternatively, the invention may also be embodied as a non-transitory machine-readable storage medium (or computer-readable storage medium, or machine-readable storage medium) having stored thereon executable code (or a computer program, or computer instruction code) which, when executed by a processor of an electronic device (or computing device, server, etc.), causes the processor to perform the steps of the above-described method according to the invention.
It should be understood that the terms "first," "second," "third," "fourth," and the like in this disclosure are used for distinguishing and not for limiting the order or importance, etc.
Those of skill would further appreciate that the various illustrative logical blocks, modules, circuits, and algorithm steps described in connection with the disclosure herein may be implemented as electronic hardware, computer software, or combinations of both.
The flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems and methods according to various embodiments of the present invention. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
Having described embodiments of the present invention, the foregoing description is intended to be exemplary, not exhaustive, and not limited to the embodiments disclosed. Many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope and spirit of the described embodiments. The terminology used herein is chosen in order to best explain the principles of the embodiments, the practical application, or improvements made to the technology in the marketplace, or to enable others of ordinary skill in the art to understand the embodiments disclosed herein.

Claims (29)

1. An image processing method for processing an image sensed by an image sensor to be displayed on a display screen, wherein the number of pixels of sensor image data sensed by the image sensor is mxn, the number of pixels of image data supplied to the display screen or a device to which the display screen belongs is rxs, R and M are the number of pixel rows, respectively, and R < M, S and N are the number of pixel columns, respectively, and S < N, the method comprising:
reading first image data corresponding to the first area from the image sensor, wherein the number of pixels of the first image data is M1 multiplied by N1, M1 is the number of pixel rows, R is not less than M1 and not more than M, N1 is the number of pixel columns, and S is not less than N1 and not more than N;
converting the first image data into second image data, wherein the pixel number of the second image data is R multiplied by S;
sending the second image data to a display screen or equipment to which the display screen belongs so as to display images;
reading third image data corresponding to the second area from the image sensor in response to a zoom instruction, the third image data having a pixel count of M2 × N2, M2 being a pixel row number, and R ≦ M2 ≦ M, N2 being a pixel column number, and S ≦ N2 ≦ N, M2 ≠ M1, N2 ≠ N1;
converting the third image data into fourth image data, wherein the pixel number of the fourth image data is R multiplied by S; and
and sending the fourth image data to the display screen or the equipment to which the display screen belongs so as to display the image.
2. An image processing method for processing an image sensed by an image sensor to be displayed on a display screen, wherein the number of pixels of sensor image data sensed by the image sensor is mxn, the number of pixels of image data supplied to the display screen or a device to which the display screen belongs is rxs, R and M are the number of pixel rows, respectively, and R < M, S and N are the number of pixel columns, respectively, and S < N, the method comprising:
reading sensor image data from an image sensor and storing the sensor image data in a memory;
extracting first image data corresponding to the first region from the sensor image data on the memory, wherein the number of pixels of the first image data is M1 Xn 1, M1 is the number of pixel rows, R is not less than M1 and not more than M, N1 is the number of pixel columns, and S is not less than N1 and not more than N;
converting the first image data into second image data, wherein the pixel number of the second image data is R multiplied by S;
sending the second image data to a display screen or equipment to which the display screen belongs so as to display images;
in response to a zoom instruction, extracting third image data corresponding to the second area from the sensor image data on the memory, the third image data having a number of pixels of M2 × N2, M2 being a number of pixel rows, and R ≦ M2 ≦ M, N2 being a number of pixel columns, and S ≦ N2 ≦ N, M2 ≠ M1, N2 ≠ N1;
converting the third image data into fourth image data, wherein the pixel number of the fourth image data is R multiplied by S; and
and sending the fourth image data to the display screen or the equipment to which the display screen belongs so as to display the image.
3. The method of claim 1 or 2,
the center of the second area coincides with the center of the first area; or
The second area is a designated area in the first area or an area where a designated object is located.
4. The method of claim 3, further comprising:
and determining the center of the second area according to the zooming instruction.
5. The method of claim 1 or 2,
the second region is a portion of the first region,
in response to the zoom instruction, before sending the fourth image data to the display screen or the device to which the display screen belongs for image display, the method further includes:
a dynamic process from displaying the image corresponding to the first area to displaying the image corresponding to the second area is presented on the display screen by a digital zoom process based on the second image data.
6. The method of claim 5, further comprising:
and responding to the display screen or the equipment to which the display screen belongs to receive the fourth image data, ending the dynamic change process display, and displaying the image based on the fourth image data.
7. The method of claim 1 or 2, further comprising:
receiving a zooming instruction from equipment and/or a server to which a display screen belongs; and/or identifying a designated object from the second image data, and issuing the zoom instruction in response to identifying the designated object in the second image data, the second region being a region in which the designated object is located.
8. An image processing method comprising:
identifying a specified object from the image data in response to the interactive request; and
in a case where the screen occupation ratio of the designated object in the image data is smaller than a predetermined occupation ratio threshold, the zoom processing is performed to increase the screen occupation ratio of the designated object in the image data after the zoom processing.
9. The method of claim 8, wherein,
the specified object is a face image,
a face recognition function is used to recognize the specified object from the second image data.
10. The method of claim 9, further comprising:
the face recognition function is turned on or off in response to the user's setting.
11. An image processing method for processing an image sensed by an image sensor to be displayed on a display screen, wherein the number of pixels of sensor image data sensed by the image sensor is mxn, the number of pixels of image data supplied to the display screen is rxs, R and M are the number of pixel rows, respectively, and R < M, S and N are the number of pixel columns, respectively, and S < N, the method comprising:
performing image conversion to obtain first image data corresponding to a first area on the image sensor and second image data for displaying on a display screen, wherein the pixel number of the first image data is M1 × N1, M1 is the pixel row number, R is not less than M1 and not more than M, N1 is the pixel column number, S is not less than N1 and not more than N, and the pixel number of the second image data is R × S;
when the distance from the specified object to the imaging device to which the image sensor belongs is greater than or equal to a first distance threshold value, sending a zooming instruction; and
and responding to a zoom instruction, performing image conversion to obtain third image data corresponding to a second area on the image sensor and fourth image data for displaying on the display screen, wherein the second area is an area where the specified object is located, the number of pixels of the third image data is M2 × N2, M2 is the number of pixel rows, R is not less than M2 and not more than M, N2 is the number of pixel columns, S is not less than N2 and not more than N, M2 is not equal to M1, N2 is not equal to N1, and the number of pixels of the fourth image data is R × S.
12. The method of claim 11, further comprising:
after the fourth image data is transmitted to the display screen or the device to which the display screen belongs for image display, in response to the distance from the specified object to the imaging device to which the image sensor belongs becoming smaller than the second distance threshold, the second image data is retransmitted to the display screen or the device to which the display screen belongs for image display.
13. An image processing method for processing an image sensed by an image sensor to be displayed on a display screen, wherein the number of pixels of sensor image data sensed by the image sensor is mxn, the number of pixels of image data supplied to the display screen is rxs, R and M are the number of pixel rows, respectively, and R < M, S and N are the number of pixel columns, respectively, and S < N, the method comprising:
performing image conversion to obtain first image data corresponding to a first area on the image sensor and second image data for displaying on a display screen, wherein the pixel number of the first image data is M1 × N1, M1 is the pixel row number, R is not less than M1 and not more than M, N1 is the pixel column number, S is not less than N1 and not more than N, and the pixel number of the second image data is R × S;
when the frame occupation ratio of the designated object relative to the first area is less than or equal to a first occupation ratio threshold value, sending a zooming instruction; and
and responding to a zoom instruction, performing image conversion to obtain third image data corresponding to a second area on the image sensor and fourth image data for displaying on the display screen, wherein the second area is an area where the specified object is located, the number of pixels of the third image data is M2 × N2, M2 is the number of pixel rows, R is not less than M2 and not more than M, N2 is the number of pixel columns, S is not less than N2 and not more than N, M2 is not equal to M1, N2 is not equal to N1, and the number of pixels of the fourth image data is R × S.
14. The method of claim 13, further comprising:
after the fourth image data is transmitted to the display screen or the device to which the display screen belongs for image display, in response to the screen occupation ratio of the specified object with respect to the first area becoming larger than the second occupation ratio threshold, the second image data is transmitted again to the display screen or the device to which the display screen belongs for image display.
15. The method of claim 13, further comprising:
the number of pixels m2 × n2 of the third image data is determined based on the screen aspect ratio of the specified object with respect to the first area and the number of pixels m1 × n1 of the first image data.
16. An image processing method for processing an image sensed by an image sensor to be displayed on a display screen, wherein the number of pixels of sensor image data sensed by the image sensor is mxn, the number of pixels of image data supplied to the display screen is rxs, R and M are the number of pixel rows, respectively, and R < M, S and N are the number of pixel columns, respectively, and S < N, the method comprising:
performing image conversion to obtain first image data corresponding to a first area on the image sensor and second image data for displaying on a display screen, wherein the pixel number of the first image data is M1 × N1, M1 is the pixel row number, R is not less than M1 and not more than M, N1 is the pixel column number, S is not less than N1 and not more than N, and the pixel number of the second image data is R × S;
identifying a specified object from the second image data;
responding to a zoom instruction, performing image conversion to obtain third image data corresponding to a second area on the image sensor and fourth image data for displaying on a display screen, wherein the second area is an area where a specified object is located, the number of pixels of the third image data is M2 × N2, M2 is the number of pixel rows, R is not less than M2 and not more than M, N2 is the number of pixel columns, S is not less than N2 and not more than N, M2 is not equal to M1, N2 is not equal to N1, and the number of pixels of the fourth image data is R × S; and
the second area is moved in response to the movement of the specified object.
17. The method of claim 16, further comprising:
in response to a partial or full movement of the second region beyond the range of the first region, expanding and/or moving the range of the first region on the image sensor such that the expanded and/or moved first region includes the second region; and/or
And in response to the part or the whole of the second area moving beyond the range of the first area, rotating the shooting direction of the imaging device to which the image sensor belongs, so that the first area on the rotated image sensor comprises the second area.
18. An image processing method for processing an image sensed by an image sensor to be displayed on a display screen, wherein the number of pixels of sensor image data sensed by the image sensor is mxn, the number of pixels of image data supplied to the display screen is rxs, R and M are the number of pixel rows, respectively, and R < M, S and N are the number of pixel columns, respectively, and S < N, the method comprising:
performing image conversion to obtain first image data corresponding to a first area on the image sensor and second image data for displaying on a display screen, wherein the pixel number of the first image data is M1 × N1, M1 is the pixel row number, R is not less than M1 and not more than M, N1 is the pixel column number, S is not less than N1 and not more than N, and the pixel number of the second image data is R × S;
identifying a specified object from the second image data;
performing image conversion to obtain third image data corresponding to a second area on the image sensor and fourth image data used for displaying on the display screen, wherein the second area is an area where a specified object is located, the number of pixels of the third image data is M2 × N2, M2 is the number of pixel rows, R is not less than M2 and not more than M, N2 is the number of pixel columns, S is not less than N2 and not more than N, M2 is not equal to M1, N2 is not equal to N1, and the number of pixels of the fourth image data is R × S; and
based on the fourth image data, the identity information of the specified object is recognized using a face recognition function.
19. The method of claim 18, further comprising:
based on the identified identity information, granting the authority of the user corresponding to the specified object to execute the predetermined operation; and/or
And performing security check on the user corresponding to the specified object based on the identified identity information.
20. An image processing apparatus for processing an image sensed by an image sensor to be displayed on a display screen, wherein the image data sensed by the image sensor has a pixel count of M × N, the image data supplied to the display screen or a device to which the display screen belongs has a pixel count of R × S, R and M are the pixel row numbers, respectively, and R < M, S and N are the pixel column numbers, respectively, and S < N, the method comprising:
first reading means for reading first image data corresponding to the first region from the image sensor, the number of pixels of the first image data being M1 × N1, M1 being the number of pixel rows, and R ≦ M1 ≦ M, N1 being the number of pixel columns, and S ≦ N1 ≦ N;
a first conversion means for converting the first image data into second image data having a pixel number of R × S;
the first sending device is used for sending the second image data to the display screen or the equipment to which the display screen belongs so as to display the image;
second reading means for reading third image data corresponding to the second area from the image sensor in response to a zoom instruction, the third image data having a pixel count of M2 × N2, M2 being a pixel row number, and R ≦ M2 ≦ M, N2 being a pixel column number, and S ≦ N2 ≦ N, M2 ≠ M1, N2 ≠ N1;
second conversion means for converting the third image data into fourth image data having a pixel count of R × S; and
and the second sending device is used for sending the fourth image data to the display screen or the equipment to which the display screen belongs so as to display the image.
21. An image processing apparatus for processing an image sensed by an image sensor to be displayed on a display screen, wherein the number of pixels of sensor image data sensed by the image sensor is mxn, the number of pixels of image data transmitted to the display screen or a device to which the display screen belongs is rxs, R and M are the number of pixel rows, respectively, and R < M, S and N are the number of pixel columns, respectively, and S < N, the method comprising:
reading means for reading sensor image data from the image sensor and storing the sensor image data in the memory;
first extracting means for extracting first image data corresponding to the first area from the sensor image data on the memory, the number of pixels of the first image data being M1 × N1, M1 being the number of pixel rows, and R ≦ M1 ≦ M, N1 being the number of pixel columns, and S ≦ N1 ≦ N;
a first conversion means for converting the first image data into second image data having a pixel number of R × S;
the first sending device is used for sending the second image data to the display screen or the equipment to which the display screen belongs so as to display the image;
second extracting means for extracting, in response to a zoom instruction, third image data corresponding to the second region from the sensor image data on the memory, the third image data having a pixel number of M2 × N2, M2 being a pixel row number, R ≦ M2 ≦ M, N2 being a pixel column number, and S ≦ N2 ≦ N, M2 ≠ M1, N2 ≠ N1;
second conversion means for converting the third image data into fourth image data having a pixel count of R × S; and
and the second sending device is used for sending the fourth image data to the display screen or the equipment to which the display screen belongs so as to display the image.
22. An image processing apparatus comprising:
the device comprises a camera device, a data acquisition device and a data processing device, wherein the camera device is used for sensing a field image on an image sensor in real time, the pixel number of sensor image data sensed by the image sensor is M multiplied by N, M is the pixel row number, and N is the pixel column number;
a display screen for displaying a live image sensed on the image sensor, wherein the number of pixels of image data provided to the display screen is R × S, R is the number of pixel rows, R < M, S is the number of pixel columns, and S < N; and
a processor for processing image data from the image sensor for display on a display screen,
the processor reads first image data corresponding to the first area from the image sensor, converts the first image data into second image data and sends the second image data to the display screen for image display, wherein the number of pixels of the first image data is M1 multiplied by N1, the number of pixels of the second image data is R multiplied by S, M1 is the number of pixel rows, R is not less than M1 and not more than M, N1 is the number of pixel columns, and S is not less than N1 and not more than N;
in response to a zoom instruction, the processor reads third image data corresponding to the second area from the image sensor, converts the third image data into fourth image data, and sends the fourth image data to the display screen for image display, wherein the number of pixels of the third image data is M2 × N2, M2 is the number of pixel rows, R ≦ M2 ≦ M, N2 is the number of pixel columns, S ≦ N2 ≦ N, M2 ≠ M1, N2 ≠ N1, and the number of pixels of the fourth image data is R × S.
23. An image processing apparatus comprising:
the device comprises a camera device, a data acquisition device and a data processing device, wherein the camera device is used for sensing a field image on an image sensor in real time, the pixel number of sensor image data sensed by the image sensor is M multiplied by N, M is the pixel row number, and N is the pixel column number;
a display screen for displaying a live image sensed on the image sensor, wherein the number of pixels of image data provided to the display screen is R × S, R is the number of pixel rows, R < M, S is the number of pixel columns, and S < N;
a memory for storing sensor image data read from the image sensor; and
a processor for processing the sensor image data for display on a display screen,
the processor extracts first image data corresponding to the first area from the sensor image data on the memory, converts the first image data into second image data and sends the second image data to the display screen for image display, wherein the number of pixels of the first image data is M1 multiplied by N1, the number of pixels of the second image data is R multiplied by S, M1 is the number of pixel rows, R is not less than M1 and not more than M, N1 is the number of pixel columns, and S is not less than N1 and not more than N;
in response to a zoom instruction, the processor extracts third image data corresponding to the second region from the sensor image data on the memory, converts the third image data into fourth image data, and sends the fourth image data to the display screen for image display, wherein the number of pixels of the third image data is M2 × N2, M2 is the number of pixel rows, and R ≦ M2 ≦ M, N2 is the number of pixel columns, and S ≦ N2 ≦ N, M2 ≠ M1, N2 ≠ N1, and the number of pixels of the fourth image data is R × S.
24. An image processing apparatus comprising:
the device comprises a camera device, a data acquisition device and a data processing device, wherein the camera device is used for sensing a field image on an image sensor in real time, the pixel number of sensor image data sensed by the image sensor is M multiplied by N, M is the pixel row number, and N is the pixel column number;
a processor for processing image data from the image sensor for display on a display screen; and
a network module for connecting to a network,
the processor reads first image data corresponding to the first area from the image sensor, converts the first image data into second image data, and sends the second image data to an external device with a display screen through the network module so as to display images, wherein the number of pixels of the first image data is M1 multiplied by N1, the number of pixels of the second image data is R multiplied by S, M1 and R are pixel row numbers respectively, R is not less than M1 and not more than M, N1 and S are pixel column numbers respectively, and S is not less than N1 and not more than N;
in response to the zoom instruction, the processor reads third image data corresponding to the second region from the image sensor, converts the third image data into fourth image data, and transmits the fourth image data to an external device having a display screen via the network module for image display, wherein the number of pixels of the third image data is M2 × N2, M2 is the number of pixel rows, and R ≦ M2 ≦ M, N2 is the number of pixel columns, and S ≦ N2 ≦ N, M2 ≠ M1, N2 ≠ N1, and the number of pixels of the fourth image data is R × S.
25. An image processing apparatus comprising:
the device comprises a camera device, a data acquisition device and a data processing device, wherein the camera device is used for sensing a field image on an image sensor in real time, the pixel number of sensor image data sensed by the image sensor is M multiplied by N, M is the pixel row number, and N is the pixel column number;
a memory for storing sensor image data read from the image sensor;
a processor for processing the sensor image data for display on a display screen; and
a network module for connecting to a network,
the processor extracts first image data corresponding to the first area from the sensor image data on the memory, converts the first image data into second image data, and sends the second image data to an external device with a display screen through the network module so as to display images, wherein the number of pixels of the first image data is M1 multiplied by N1, the number of pixels of the second image data is R multiplied by S, M1 and R are pixel row numbers respectively, R is not less than M1 and not more than M, N1 and S are pixel column numbers respectively, and S is not less than N1 and not more than N;
in response to a zoom instruction, the processor extracts third image data corresponding to the second area from the sensor image data on the memory, converts the third image data into fourth image data, and transmits the fourth image data to an external device having a display screen via the network module for image display, wherein the third image data has a pixel count of M2 × N2, M2 is a pixel row count, and R ≦ M2 ≦ M, N2 is a pixel column count, and S ≦ N2 ≦ N, M2 ≠ M1, N2 ≠ N1, and the fourth image data has a pixel count of R × S.
26. An image processing system comprising:
the image processing apparatus according to claim 24 or 25; and
and the electronic equipment is provided with a display screen, receives the second image data and the fourth image data from the image processing equipment and displays images correspondingly.
27. The system of claim 26, wherein,
the image processing apparatus further comprises a local display screen,
the image processing device receives an image from the electronic device via a network module and displays the received image on a local display screen.
28. A computing device, comprising:
a processor; and
a memory having executable code stored thereon, which when executed by the processor, causes the processor to perform the method of any of claims 1 to 19.
29. A non-transitory machine-readable storage medium having stored thereon executable code, which when executed by a processor of an electronic device, causes the processor to perform the method of any of claims 1-19.
CN202010389013.8A 2020-05-09 2020-05-09 Image processing method, device and system Pending CN113630546A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010389013.8A CN113630546A (en) 2020-05-09 2020-05-09 Image processing method, device and system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010389013.8A CN113630546A (en) 2020-05-09 2020-05-09 Image processing method, device and system

Publications (1)

Publication Number Publication Date
CN113630546A true CN113630546A (en) 2021-11-09

Family

ID=78377640

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010389013.8A Pending CN113630546A (en) 2020-05-09 2020-05-09 Image processing method, device and system

Country Status (1)

Country Link
CN (1) CN113630546A (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2006094082A (en) * 2004-09-24 2006-04-06 Casio Comput Co Ltd Image photographing device, and program
CN102202181A (en) * 2007-04-02 2011-09-28 捷讯研究有限公司 Camera with multiple viewfinders
CN107896303A (en) * 2017-10-23 2018-04-10 努比亚技术有限公司 A kind of image-pickup method, system and equipment and computer-readable recording medium
CN109784256A (en) * 2019-01-07 2019-05-21 腾讯科技(深圳)有限公司 Face identification method and device, storage medium and electronic device

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2006094082A (en) * 2004-09-24 2006-04-06 Casio Comput Co Ltd Image photographing device, and program
CN102202181A (en) * 2007-04-02 2011-09-28 捷讯研究有限公司 Camera with multiple viewfinders
CN107896303A (en) * 2017-10-23 2018-04-10 努比亚技术有限公司 A kind of image-pickup method, system and equipment and computer-readable recording medium
CN109784256A (en) * 2019-01-07 2019-05-21 腾讯科技(深圳)有限公司 Face identification method and device, storage medium and electronic device

Similar Documents

Publication Publication Date Title
KR102653850B1 (en) Digital photographing apparatus and the operating method for the same
KR102458339B1 (en) Electronic Apparatus generating 360 Degrees 3D Stereoscopic Panorama Images and Method thereof
US11758265B2 (en) Image processing method and mobile terminal
US10437545B2 (en) Apparatus, system, and method for controlling display, and recording medium
CN101689292B (en) Banana codec
US10748243B2 (en) Image distortion transformation method and apparatus
US20130265311A1 (en) Apparatus and method for improving quality of enlarged image
KR102492670B1 (en) Image processing apparatus, image processing method, and storage medium
KR101530255B1 (en) Cctv system having auto tracking function of moving target
JP6727989B2 (en) Image processing apparatus and control method thereof
US20140204083A1 (en) Systems and methods for real-time distortion processing
US20150262330A1 (en) Image display apparatus and image display method
CN112017137A (en) Image processing method, image processing device, electronic equipment and computer readable storage medium
US20210065331A1 (en) Image processing apparatus, image communication system, image processing method, and recording medium
CN105467741B (en) A kind of panorama photographic method and terminal
JP5805013B2 (en) Captured image display device, captured image display method, and program
WO2017024954A1 (en) Method and device for image display
CN113630546A (en) Image processing method, device and system
US20100302403A1 (en) Generating Images With Different Fields Of View
KR20210080334A (en) Method, apparatus, and device for identifying human body and computer readable storage
US10943562B2 (en) Display control apparatus and display control method
RU2792413C1 (en) Image processing method and mobile terminal
JP6979799B2 (en) How to shoot camera and video
JP7244661B2 (en) Imaging device and image processing method
US20150116313A1 (en) Information processing apparatus, information processing method, and program

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20211109