CN110248004A - Terminal, the control method of terminal and image acquiring method - Google Patents
Terminal, the control method of terminal and image acquiring method Download PDFInfo
- Publication number
- CN110248004A CN110248004A CN201910556578.8A CN201910556578A CN110248004A CN 110248004 A CN110248004 A CN 110248004A CN 201910556578 A CN201910556578 A CN 201910556578A CN 110248004 A CN110248004 A CN 110248004A
- Authority
- CN
- China
- Prior art keywords
- viewing area
- imaging device
- algorithm
- terminal
- display screen
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/02—Constructional features of telephone sets
- H04M1/0202—Portable telephone sets, e.g. cordless phones, mobile phones or bar type handsets
- H04M1/026—Details of the structure or mounting of specific components
- H04M1/0264—Details of the structure or mounting of specific components for a camera module assembly
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/02—Constructional features of telephone sets
- H04M1/0202—Portable telephone sets, e.g. cordless phones, mobile phones or bar type handsets
- H04M1/026—Details of the structure or mounting of specific components
- H04M1/0266—Details of the structure or mounting of specific components for a display module assembly
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/57—Mechanical or electrical details of cameras or camera modules specially adapted for being embedded in other devices
Landscapes
- Engineering & Computer Science (AREA)
- Signal Processing (AREA)
- Multimedia (AREA)
- Controls And Circuits For Display Device (AREA)
- Studio Devices (AREA)
Abstract
This application discloses a kind of terminals.Terminal includes display screen and imaging device, and display screen includes the first viewing area and the second viewing area;Imaging device is correspondingly arranged with the first viewing area, and imaging device is for receiving the optical signal across the first viewing area to be imaged;When imaging device is imaged, the first viewing area is in OFF state.Disclosed herein as well is a kind of control method of terminal and image acquiring methods.First viewing area of imaging device and display screen is correspondingly arranged, and imaging device receives the optical signal across the first viewing area to be imaged, and does not need to open up the through-hole for passing through for optical signal on a display screen, and then improve the screen accounting of terminal.
Description
Technical field
This application involves consumer electronics technical fields, more specifically, be related to a kind of terminal, the control method of terminal and
Image acquiring method.
Background technique
With the fast development of smart phone, screen becomes the basic demand to mobile phone comprehensively, improves screen accounting, is consumer
The trend for the development for bringing the visual experience of Rimless large-size screen monitors to become mobile phone industry is provided.But this is needed due to mobile phone self-timer
It asks, front camera just needs to be placed in front, it usually needs form through-hole on the screen to be used to place front camera, so
And forming through-hole causes the screen accounting of mobile phone lower.
Summary of the invention
The application embodiment provides a kind of terminal, the control method of terminal and image acquiring method.
A kind of terminal that the application embodiment provides includes display screen and imaging device, and the display screen includes first aobvious
Show area and the second viewing area;The imaging device is correspondingly arranged with first viewing area, and the imaging device is worn for receiving
The optical signal of first viewing area is crossed to be imaged;When the imaging device is imaged, first viewing area is in
OFF state.
Another terminal that the application embodiment provides includes display screen, imaging device and processor, the display screen
Including the first viewing area and the second viewing area;The imaging device is correspondingly arranged with first viewing area, the imaging device
For receiving the optical signal across first viewing area to obtain initial pictures;The processor is for handling the initial graph
As to obtain final image.
In some embodiments, when the imaging device obtains the initial pictures, first viewing area, which is in, puts out
It going out state, the processor is used to correct the initial pictures by the first algorithm to obtain the final image, and described first
Algorithm is related to the screen structure of first viewing area.
In some embodiments, the quantity of first algorithm is multiple, each first algorithm and an imaging
Scene is corresponding, and the processor is also used to: obtaining the first algorithm corresponding with the image scene of the initial pictures;And by pair
First algorithm answered corrects the initial pictures to obtain the final image.
In some embodiments, in calibration, first viewing area is in OFF state, and the imaging device is used for
The optical signal across first viewing area is received to obtain test image, and receives the light letter for not passing through first viewing area
Number to obtain reference picture;The processor is used to obtain described first according to the test image and the reference picture and calculate
Method.
In some embodiments, first algorithm includes luminance compensation algorithm, diffraction pattern elimination algorithm, completion calculation
One of method and defogging algorithm are a variety of.
In some embodiments, when the imaging device obtains the initial pictures, first viewing area is shown in advance
If picture, the processor is used to correct the initial pictures by the second algorithm to obtain the final image, described second
Algorithm is related to the screen structure of first viewing area and the default picture.
In some embodiments, the display screen includes the first son screen and the second son screen that can be independently controlled, institute
It states the first viewing area to be formed on the first son screen, second viewing area is formed on the second son screen;Or
The display screen is an independent screen structure, and first viewing area and second viewing area can be by independences
Control.
In some embodiments, the non-corner region of the display screen is arranged in first viewing area.
In some embodiments, the interposition at least one edge of the display screen is arranged in first viewing area
It sets.
In some embodiments, the pixel density of first viewing area is close less than the pixel of second viewing area
Degree.
In some embodiments, when the imaging device is not imaged, first viewing area is aobvious with described second
Show that area is in illuminating state.
The application embodiment provide terminal control method be used for controlling terminal, the terminal include display screen and at
As device, the display screen includes the first viewing area and the second viewing area, and the imaging device is corresponding with first viewing area
Setting;The control method includes: that control first viewing area is in OFF state;And the control imaging device reception is worn
The optical signal of first viewing area is crossed to be imaged.
A kind of image acquiring method that the application embodiment provides is used for terminal, and the terminal includes display screen and imaging
Device, the display screen include the first viewing area and the second viewing area, and the imaging device is corresponding with first viewing area to be set
It sets, described image acquisition methods include: that control first viewing area is in OFF state;The imaging device reception is controlled to wear
The optical signal of first viewing area is crossed to obtain initial pictures;And the initial pictures are corrected to obtain most by the first algorithm
Whole image, first algorithm are related to the screen structure of first viewing area.
The application embodiment provide another image acquiring method be used for terminal, the terminal include display screen and at
As device, the display screen includes the first viewing area and the second viewing area, and the imaging device is corresponding with first viewing area
Setting, described image acquisition methods include: that control first viewing area shows default picture;The imaging device is controlled to receive
Optical signal across first viewing area is to obtain initial pictures;And the initial pictures are corrected to obtain by the second algorithm
Final image, second algorithm are related to the screen structure of first viewing area and the default picture.
In the terminal of the application embodiment, the control method of terminal and image acquiring method, imaging device and display screen
The first viewing area be correspondingly arranged, imaging device receives the optical signal across the first viewing area to be imaged, and does not need aobvious
The through-hole for passing through for optical signal is opened up in display screen, and then improves the screen accounting of terminal.
The additional aspect and advantage of presently filed embodiment will be set forth in part in the description, partially will be from following
Description in become obvious, or recognized by the practice of presently filed embodiment.
Detailed description of the invention
The above-mentioned and/or additional aspect and advantage of the application is from combining in description of the following accompanying drawings to embodiment by change
It obtains obviously and is readily appreciated that, in which:
Fig. 1 is the structural schematic diagram of the terminal of the application embodiment;
Fig. 2 is schematic cross-section of the terminal shown in FIG. 1 along II-II line;
Fig. 3 and Fig. 4 is the structural schematic diagram of the terminal of the application embodiment;
Fig. 5 is the structural schematic diagram of the display screen of the application embodiment;
Fig. 6 is the part-structure enlarged diagram of the terminal of the application embodiment;
Fig. 7 is the structural schematic diagram of the sensor of the imaging device of the application embodiment;
Fig. 8 is the calibration schematic diagram of a scenario of the terminal of the application embodiment;
Fig. 9 is the flow diagram of the control method of the terminal of the application embodiment;
Figure 10 and Figure 11 is the flow diagram of the image acquiring method of the application embodiment.
Main element symbol description:
Terminal 100, casing 10, accommodating chamber 11, display screen 20, the first viewing area 21, second viewing area the 22, first son screen
23, the second son screen 24, pixel 25, imaging device 30, sensor 31, photosensitive unit 311, imaging unit 312, cover board 40, processing
Device 50.
Specific embodiment
Presently filed embodiment is described further below in conjunction with attached drawing.Same or similar label is from beginning in attached drawing
To the same or similar element of expression or element with the same or similar functions eventually.
In addition, the presently filed embodiment described with reference to the accompanying drawing is exemplary, it is only used for explaining the application's
Embodiment, and should not be understood as the limitation to the application.
In this application unless specifically defined or limited otherwise, fisrt feature in the second feature " on " or " down " can be with
It is that the first and second features directly contact or the first and second features pass through intermediary mediate contact.Moreover, fisrt feature exists
Second feature " on ", " top " and " above " but fisrt feature be directly above or diagonally above the second feature, or be merely representative of
First feature horizontal height is higher than second feature.Fisrt feature can be under the second feature " below ", " below " and " below "
One feature is directly under or diagonally below the second feature, or is merely representative of first feature horizontal height less than second feature.
Fig. 1 and Fig. 2 is please referred to, the terminal 100 of the application embodiment includes display screen 20 and imaging device 30.Display screen
20 include the first viewing area 21 and the second viewing area 22.Imaging device 30 is correspondingly arranged with the first viewing area 21, imaging device 30
For receiving the optical signal across the first viewing area 21 to be imaged.
In above-mentioned terminal 100, since imaging device 30 and the first viewing area 21 of display screen 20 are correspondingly arranged, imaging device
30 receive the optical signal across the first viewing area 21 to be imaged, and do not need to be opened up on display screen 20 for wearing for optical signal
The through-hole crossed, and then improve the screen accounting of terminal 100.
Specifically, 1 and Fig. 2 is please referred to, the terminal 100 of the application embodiment includes casing 10, display screen 20, imaging dress
Set 30 and processor 50.The concrete form of terminal 100 can be mobile phone, game machine, tablet computer, display, laptop,
Smartwatch, head show equipment etc., and the application is illustrated so that terminal 100 is mobile phone as an example, it will be understood that terminal 100 it is specific
Form is not limited to mobile phone, and this is not restricted.
Casing 10 can be the shell of terminal 100, the e.g. center and rear cover of mobile phone.Casing 10 can be used for installing aobvious
The function elements such as display screen 20, imaging device 30, processor 50, function element can also include mainboard, power module etc..Terminal
100 may also include cover board 40, and cover board 40 can be made of the transparent material such as glass, sapphire, resin, can collect on cover board 40
At there is touch-control sensing layer, to incude the touch operation on cover board 40.Cover board 40 and casing 10 can surround accommodating chamber 11 jointly,
Display screen 20, imaging device 30 and processor 50 can be housed in accommodating chamber 11 so that display screen 20, imaging device 30 and
Processor 50 is not easily susceptible to the erosion of steam and dust etc..
Please continue to refer to Fig. 1 and Fig. 2, display screen 20 is mounted on casing 10, and display screen 20 can be used for issuing optical signal, light
Signal enters outside terminal 100 after passing through cover board 40, so that display screen 20 shows the pictures such as picture, video, text.Specifically,
Display screen 20 may be mounted on a face of casing 10, or is mounted on two faces of casing 10 simultaneously, such as exist simultaneously
Display screen 20 is installed on the front and the back side of casing 10, or is mounted on the two or more face of casing 10 simultaneously, such as together
When on the front, the back side and side of casing 10 install display screen 20.In example as shown in Figure 1, display screen 20 is mounted on
The front of casing 10, display screen 20 can cover 90% or more of the positive area, for example, can be 91%, 92%,
93%, 95%, 97%, even 100%.The global shape of display screen 20 can be rectangle, circle, ellipse, track type, circle
The shapes such as angular moment shape, triangle, this is not restricted.
Display screen 20 includes the first viewing area 21 and the second viewing area 22.First viewing area 21 and the second viewing area 22 it is aobvious
Show that state can be independently controlled, for example, the first viewing area 21 and the second viewing area 22 extinguishing, light, show content, display
Brightness, refreshing frequency etc. can be independently controlled.Wherein, extinguish and refer to viewing area (the first viewing area 21 or the second viewing area 22)
In the state not shown, the state for referring to that viewing area (the first viewing area 21 or the second viewing area 22) is in display is lighted.One
In a example, the first viewing area 21 and the second viewing area 22 can show the continuous picture of a frame, such as the first viewing area jointly
21 and second viewing area 22 show continuous two parts of the same object respectively.In another example, the first viewing area 21
It can show that different pictures, such as the second viewing area 22 show current picture, the first viewing area respectively from the second viewing area 22
21 show that the thumbnail of the current picture or the first viewing area 21 can be used for the battery capacity of display terminal 100, network connection
The Status icons such as state, system time.
Fig. 1, Fig. 3 and Fig. 4 are please referred to, the first viewing area 21 can be set in the non-corner region of display screen 20.Wherein, angle
Portion region refers to the partial region of two edges intersection of display screen 20.It is appreciated that terminal 100 is when dropping, display screen 20
Folding corner region is particularly susceptible rupture, is easy to be damaged if folding corner region, the first viewing area 21 if being arranged the first viewing area 21,
And the imaging device 30 being correspondingly arranged with the first viewing area 21 is also easy to happen displacement or damage.And the first viewing area 21 is set
It sets a possibility that non-corner region can reduce the first viewing area 21 to a certain extent and imaging device 30 damages.
Specifically, the first viewing area 21 can be set in the middle position at least one edge of display screen 20, to use
When carrying out self-timer using imaging device 30, user can also look at imaging device 30 straight when looking at display screen 20 straight simultaneously, self-timer at family
Effect is preferable.Such as in example as shown in Figures 1 and 3, the first viewing area 21 is arranged in the top edge of display screen 20
Between position;In the example as depicted in fig. 4, the middle position of the lower edge of display screen 20 is arranged in the first viewing area 21.Certainly,
In other examples, the middle position of the left edge of display screen 20, right hand edge etc. can also be arranged in the first viewing area 21, or
It is arranged on the middle position of two edges (such as top edge and lower edge) simultaneously, this is not restricted.It should be noted that
In the attached drawing of this specification, the boundary line of the first viewing area 21 and the second viewing area 22 is intended merely to clear explanation boundary between the two
Limit, does not represent and will appear the boundary line in actual displayed.
The shape of first viewing area 21 and the second viewing area 22 can be configured according to demand, for example, the first viewing area
21 can be the arbitrary shapes such as rectangle, track type, water-drop-shaped, semicircle, circle, and the second viewing area 22 can be with the first viewing area
21 are complementarily shaped to the global shape of above-mentioned display screen 20, for example, as shown in Figure 1, the second viewing area 22 and the first viewing area 21
It is complementarily shaped to the shape display screen 20 of round rectangle.The area of first viewing area 21 can be set less than the second viewing area 22
Area, such as the area of the first viewing area 21 is 1/10th, 1/20th, 25 points of area of the second viewing area 22
One of equal proportion relationship.
The concrete composition form of display screen 20 can have different selections, and in one example, display screen 20 is one only
Vertical screen structure, the first viewing area 21 and the second viewing area 22 can be independently controlled.At this point, the first viewing area 21 and second is aobvious
Show that area 22 is the different subregions of the independent screen structure, there is no apparent between the first viewing area 21 and the second viewing area 22
Boundary.Display screen 20 can be OLED display screen, Micro LED display or be by can independent control dot matrix backlight
The liquid crystal display of (such as mini led or micro led) as backlight.
Referring to Fig. 5, in another example, display screen 20 includes the first son screen 23 and second that can be independently controlled
Son screen 24, the first viewing area 21 are formed on the first son screen 23, and the second viewing area 22 is formed on the second son screen 24.That is, display
Screen 20 is made of two independent sub- display devices (the first son screen 23 and the second son screen 24).First son screen 23 and the second son screen 24
When being combined into display screen 20, human eye is difficult to perceive the boundary line between the first son screen 23 and the second son screen 24.First son screen
23 drive display by different driving circuits respectively from the second son screen 24.Second son screen 24 can form notch, and notch for example may be used
To be the shapes such as track type, rectangle, water-drop-shaped, the shape of the first son screen 23 can be identical as the shape of notch, so that the first son
Screen 23 and the second son screen 24 can be combined into more regular display screen 20.At this point, the first son screen 23 shield 24 with the second son can be with
The display device of same type, for example, first son screen 23 with second son shield 24 be OLED display screen, Micro LED display,
Or one of any form of display device such as liquid crystal display;First son screen 23 and the second son screen 24 are also possible to inhomogeneity
The display device of type, for example, first son screen 23 with second son shield 24 can be respectively OLED display screen, Micro LED display,
Or the different two kind display device in any form of display device such as liquid crystal display, this is not restricted.
Fig. 1 and Fig. 2 is please referred to, in the casing 10 that imaging device 30 is mounted on.Imaging device 30 and the first viewing area 21 are right
It should be arranged, imaging device 30 is for receiving the optical signal across the first viewing area 21 to be imaged.Imaging device 30 and first
Viewing area 21 is correspondingly arranged, and can be the incidence surface of imaging device 30 towards the first viewing area 21, optical signal passes through the first display
It behind area 21, can be acquired by imaging device 30 and for being imaged, in other words, imaging device 30 is set under display screen 20, and with the
The alignment of one viewing area 21.
Specifically, imaging device 30 can be what the visible light signal for passing through the first viewing area 21 for foundation was imaged
Imaging device 30, such as imaging device 30 are color camera or black and white camera etc.;Imaging device 30 can also be to be worn for foundation
The imaging device 30 that the black light signal of the first viewing area 21 is imaged is crossed, such as imaging device 30 is infrared camera or purple
Outer camera etc.;Imaging device 30 can also be the reception device in depth camera, such as imaging device 30 is structure optical depth phase
Optical receiver apparatus in machine, or be the optical receiver apparatus in flight time (Time of flight, TOF) depth camera,
This is with no restriction.Present specification is received for visible light signal is imaged by imaging device 30 and is illustrated.
When display screen 20 is mounted on the front of casing 10, imaging device 30 can be preposition imaging device;Work as display screen
20 when being mounted on the back side of casing 10, and imaging device 30 can be postposition imaging device;Certainly, when display screen 20 is mounted on simultaneously
When the front and the back side of casing 10, imaging device 30 may include preposition imaging device and postposition imaging device.The application explanation
Book is illustrated so that imaging device 30 is preposition imaging device as an example.
Imaging device 30 may include single camera, and single camera is aligned with the first viewing area 21;Imaging device 30
It also may include two cameras, two cameras are aligned with the first viewing area 21;Imaging device 30 also may include three
Camera, three cameras are aligned with the first viewing area 21;Certainly, imaging device 30 can also include more cameras,
More cameras are aligned with the first viewing area 21.
Incorporated by reference to Fig. 6, the optical signal due to imaging device 30 for imaging is needed across the first viewing area 21, and first is aobvious
Show that the screen structure in area 21 itself may have an impact optical signal, such as the pixel 25 in the first viewing area 21 may make light
Signal generates diffraction or weakens the intensity etc. of optical signal, and 30 imaging of imaging device is caused not to be inconsistent with actual scene.Cause
This, it may be considered that this influence is reduced by the structure of design display screen 20 and imaging device 30.
Specifically, for display screen 20,25 density of pixel of the first viewing area 21 can be reduced to reduce and show to first
Effect of the area 21 to optical signal, as shown in fig. 6,25 density of pixel of the first viewing area 21 can be set less than the second viewing area 22
25 density of pixel, such as the first viewing area 21 25 density of pixel be the second viewing area 22 25 density of pixel three/
One, half, a quarter etc. so that the display effect of the second viewing area 22 is preferable, while imaging device 30 receive across
The optical signal of first viewing area 21 and for imaging image quality it is preferable.
Connecting line for connecting the pixel 25 of the first viewing area 21 can be transferred to as much as possible in the second viewing area 22,
Or be transferred in the region of the first viewing area 21 not being aligned with imaging surface, to reduce the connecting line pair in the first viewing area 21
The influence of optical signal.Meanwhile first the selection of component part of viewing area 21 also can choose the higher material of light transmittance, such as
Select transparent material (such as ITO) that the connecting line of pixel 25 is made;If there is polaroid in the first viewing area 21, also may be selected transparent
Material polaroid is made.
For imaging device 30, incorporated by reference to Fig. 7, imaging device 30 can choose the higher sensor 31 of sensitivity, one
In a example, the size of the photosensitive unit 311 of sensor 31 can be increased, be 1 micron × 1 micro- for example, by using length (L) × wide (h)
The photosensitive unit 311 of rice can increase each photosensitive compared to photosensitive unit of the length × width for 0.8 micron × 0.8 micron is used
The amount for the optical signal that unit 311 can receive obtains preferable image quality.In one example, it can satisfy imaging point
In the case where resolution, multiple photosensitive units 311 in sensor 31 can be merged into an imaging unit 312 and used, this one
Multiple photosensitive units 311 in a imaging unit 312 add up to the amount of received optical signal for reflecting the single picture being ultimately imaged
The pixel value of element.The quantity of photosensitive unit 311 in one imaging unit 312 can be two, four, eight, 16 etc.
Any amount.Furthermore it is also possible to using the imaging device 30 with more large aperture camera lens, to increase light-inletting quantity, example to greatest extent
F2.0, F1.8 or bigger aperture can such as be used.
Imaging device 30 can also cooperate jointly to reduce influence of first viewing area 21 to optical signal with the first viewing area 21.
In one example, when imaging device 30 is imaged, the first viewing area 21 is in OFF state.First viewing area 21 is in
When OFF state, the first viewing area 21 will not actively launch outward optical signal, and the optical signal for avoiding the first viewing area 21 from issuing is straight
It taps into imaging device 30 and influences imaging.Specifically, when the processor 50 of terminal 100 receives the unlatching imaging device of user
When 30 instruction, processor 50 controls the first viewing area 21 and extinguishes, and opens simultaneously imaging device 30, so that imaging device 30 receives
Optical signal across the first viewing area 21 is to be imaged.At this point, the second viewing area 22 can normally be shown, such as can be used for
Show the preview screen that imaging device 30 obtains.After 30 use of imaging device, that is, imaging device 30 is not imaged
When, the first viewing area 21 and the second viewing area 22 can be in illuminating state, specifically can control first by processor 50 and show
Show that area 21 lights jointly with the second viewing area 22, to show more complete picture, improves the usage experience of user.
In another example, when imaging device 30 is imaged, the first viewing area 21 shows default picture.Specifically
Ground, when the processor 50 of terminal 100 receives the fingerprint of unlatching imaging device 30 of user, first display of the control of processor 50
Area 21 shows default picture, opens simultaneously imaging device 30, so that the light that imaging device 30 receives across the first viewing area 21 is believed
Number to be imaged.Wherein, default picture can be fixed picture, so that default influence of the picture to imaging is relatively fixed, after
Phase is easy to be corrected imaging.Default picture can be one of the default setting of terminal 100, such as default picture is to laugh at
The patterns such as face, eggplant, to assist user to carry out self-timer, default picture can also be current imaging parameters or for being arranged
The pictures such as the icon of imaging parameters.
At this point, the second viewing area 22 can normally be shown, such as it is displayed for the preview picture of the acquisition of imaging device 30
Face.After 30 use of imaging device, that is, when imaging device 30 is not imaged, the first viewing area 21 and the second viewing area
22 can be in illuminating state, specifically can control the first viewing area 21 and 22 common ground of the second viewing area by processor 50
It is bright, to show more complete picture, improve the usage experience of user.
Fig. 1 and Fig. 2 is please referred to, processor 50 can be all connected with imaging device 30 and display screen 20.Processor 50 can be used for
Control the use state of imaging device 30 and display screen 20.In addition, processor 50 can also be used to handle formed by imaging device 30
Picture.
As described above, imaging device 30 can be used for receiving the optical signal across the first viewing area 21 to be imaged, it can
It is known as initial pictures with the image that the unprocessed device 50 for obtaining imaging device 30 is handled.Processor 50 can be used for handling just
Beginning image is to obtain final image.
Due to the work of the default picture of the display of the effect and/or the first viewing area 21 of the screen structure of the first viewing area 21
With causing initial pictures might have, brightness is partially dark, clarity is lower, image sends out mist, hot spot (flare), ghost (ghost) etc.
Defect.After processor 50 further handles initial pictures, the higher final image of quality can get.Final image can be used for
It finally is presented to user or saves.
In one embodiment, when obtaining initial pictures for imaging device 30, the first viewing area 21 is in OFF state
Mode, processor 50 can correct initial pictures by the first algorithm to obtain final image, and the first algorithm is shown with first
The screen structure in area 21 is related.When obtaining initial pictures due to imaging device 30, the first viewing area 21 is in OFF state, therefore passes through
First algorithm relevant to the screen structure of the first viewing area 21 is corrected initial pictures, then available quality is preferably most
Whole image.Specifically, the first algorithm may include that luminance compensation algorithm, diffraction pattern elimination algorithm, completion algorithm and defogging are calculated
One of method is a variety of, at the beginning of realizing the luminance compensation to initial pictures, the diffraction pattern for eliminating initial pictures, completion respectively
The functions such as the part being atomized in the missing, removal initial pictures in beginning image.
The quantity of first algorithm can have multiple, and each first algorithm is corresponding with an image scene.Processor 50 is in school
When positive initial pictures, it is also used to obtain the first algorithm corresponding with the image scene of initial pictures, and calculate by corresponding first
Method corrects initial pictures to obtain final image.The properties such as power, the distribution situation of the optical signal of different image scenes are not
Together, therefore, for different image scenes, more targeted place can be carried out to initial pictures using corresponding first algorithm
Reason.Specifically, for example, for image scene be half-light scene, corresponding first algorithm can be more focused on to initial pictures into
Row luminance compensation;It is backlit scene for image scene, corresponding first algorithm, which can be more focused on, spreads out to initial pictures
Penetrate the elimination etc. of hot spot.
First algorithm can be obtained by calibration, specifically, be in and put out in the first viewing area 21 such as Fig. 8 a incorporated by reference to Fig. 8
When state of going out, imaging device 30 receives the optical signal across the first viewing area 21 to obtain test image P1, fills as Fig. 8 b is imaged
It sets 30 receptions and does not pass through the optical signal of the first viewing area 21 to obtain reference picture P2.Processor 50 is used for according to test image P1
And reference picture P2 obtains the first algorithm.Wherein, test image P1 and reference picture P2 is same imaging device 30 to same
One scene is imaged, and since the optical signal for obtaining test image P1 has passed through the first viewing area 21, for obtaining
The optical signal of reference picture P2 does not pass through the first viewing area 21, and therefore, reference picture P2 is different from the details of test image P1, and
Reference picture P2 can more embody the objective circumstances of image scene, such as the same light source is light source in test image P1
200a, is light source 200b in reference picture P2, and the clarity of light source 200b is higher than the clarity of light source 200a.Processor 50 will
Compare test image P1 and with reference to figure by the mode of artificial intelligence (Artificial Intelligence, AI) deep learning
As P2, and obtain the first algorithm.In calibration, imaging device 30 is needed to multiple (such as 10,000,100,000, million etc.)
Different scene obtains test image and reference picture, and processor 50 obtains and multiple and different scene one-to-one multiple the
One algorithm, in actual use, to use corresponding first algorithm according to current image scene.
In another embodiment, when obtaining initial pictures for imaging device 30, the first viewing area 21 shows default picture
The mode in face, processor 50 can correct initial pictures by the second algorithm to obtain final image, and the second algorithm and first show
Show that screen structure and the default picture in area 21 are related.When obtaining initial pictures due to imaging device 30, the first viewing area 21 shows pre-
If picture, therefore initial pictures are corrected by the second algorithm relevant to the screen structure of the first viewing area 21 and default picture, then
The available preferable final image of quality.Specifically, the second algorithm may include luminance compensation algorithm, diffraction pattern elimination calculation
One of method, completion algorithm and defogging algorithm are a variety of, to realize the luminance compensation to initial pictures respectively, eliminate initial graph
The diffraction pattern of picture, removes the functions such as the part being atomized in initial pictures at the missing in completion initial pictures.
The quantity of second algorithm can have multiple, and each second algorithm is corresponding with an image scene.Processor 50 is in school
When positive initial pictures, it is also used to obtain the second algorithm corresponding with the image scene of initial pictures, and calculate by corresponding second
Method corrects initial pictures to obtain final image.The properties such as power, the distribution situation of the optical signal of different image scenes are not
Together, therefore, for different image scenes, more targeted place can be carried out to initial pictures using corresponding second algorithm
Reason.Specifically, for example, for image scene be half-light scene, corresponding second algorithm can be more focused on to initial pictures into
Row luminance compensation;It is backlit scene for image scene, corresponding second algorithm, which can be more focused on, spreads out to initial pictures
Penetrate the elimination etc. of hot spot.
Second algorithm can be obtained by calibration, specifically, be in and put out in the first viewing area 21 such as Fig. 8 a incorporated by reference to Fig. 8
When state of going out, imaging device 30 receives the optical signal across the first viewing area 21 to obtain test image P1, fills as Fig. 8 b is imaged
It sets 30 receptions and does not pass through the optical signal of the first viewing area 21 to obtain reference picture P2.Processor 50 is used for according to test image P1
And reference picture P2 obtains the second algorithm.Wherein, test image P1 and reference picture P2 is same imaging device 30 to same
One scene is imaged, and since the optical signal for obtaining test image P1 has passed through the first viewing area 21, for obtaining
The optical signal of reference picture P2 does not pass through the first viewing area 21, and therefore, reference picture P2 is different from the details of test image P1, and
Reference picture P2 can more embody the objective circumstances of image scene, such as the same light source is light source in test image P1
200a, is light source 200b in reference picture P2, and the clarity of light source 200b is higher than the clarity of light source 200a.Processor 50 will
Compare test image P1 and reference picture P2 by way of artificial intelligence deep learning, and obtains the second algorithm.In calibration,
Imaging device 30 need to carry out multiple (such as 10,000,100,000, million etc.) different scene obtain test image and
Reference picture, processor 50 obtain with one-to-one multiple second algorithms of multiple and different scene, in actual use, according to
Corresponding second algorithm is used according to current image scene.
Incorporated by reference to Fig. 9, the control method of the terminal 100 of the application embodiment can be used for controlling the application embodiment
Terminal 100, control method comprising steps of
091: the first viewing area 21 of control is in OFF state;And
092: control imaging device 30 receives the optical signal across the first viewing area 21 to be imaged.
Step 091 and step 092 can be implemented by processor 50, specific implementation detail, can refer to above-mentioned to terminal
100 description, details are not described herein.
Incorporated by reference to Figure 10, the image acquiring method of the application embodiment can be used for the terminal 100 of the application embodiment,
Image acquiring method comprising steps of
0101: the first viewing area 21 of control is in OFF state;
0102: control imaging device 30 receives the optical signal across the first viewing area 21 to obtain initial pictures;And
0103: initial pictures are corrected to obtain final image by the first algorithm.
Step 0101, step 0102 and step 0103 can be implemented by processor 50, and specific implementation detail can refer to
The above-mentioned description to terminal 100, details are not described herein.
Incorporated by reference to Figure 11, the image acquiring method of the application embodiment can be used for the terminal 100 of the application embodiment,
Image acquiring method comprising steps of
0111: the first viewing area 21 of control shows default picture;
0112: control imaging device 30 receives the optical signal across the first viewing area 21 to obtain initial pictures;And
0113: initial pictures are corrected to obtain final image by the second algorithm.
Step 0111, step 0112 and step 0113 can be implemented by processor 50, and specific implementation detail can refer to
The above-mentioned description to terminal 100, details are not described herein.
To sum up, in the terminal 100 of the application embodiment, the control method of terminal 100 and image acquiring method, imaging dress
30 are set to be correspondingly arranged with the first viewing area 21 of display screen 20, imaging device 30 receive the optical signal across the first viewing area 31 with
It is imaged, does not need to open up the through-hole for passing through for optical signal on display screen 20, and then the screen for improving terminal 100 accounts for
Than.
In the description of this specification, reference term " certain embodiments ", " embodiment ", " some embodiment party
The description of formula ", " exemplary embodiment ", " example ", " specific example " or " some examples " means in conjunction with the embodiment
Or example particular features, structures, materials, or characteristics described are contained at least one embodiment or example of the application.
In the present specification, schematic expression of the above terms are not necessarily referring to identical embodiment or example.Moreover, description
Particular features, structures, materials, or characteristics can be in any one or more embodiments or example with suitable side
Formula combines.
In addition, term " first ", " second " are used for descriptive purposes only and cannot be understood as indicating or suggesting relative importance
Or implicitly indicate the quantity of indicated technical characteristic.Define " first " as a result, the feature of " second " can be expressed or
Implicitly include at least one described feature.In the description of the present application, the meaning of " plurality " is at least two, such as two,
Three, unless otherwise specifically defined.
Although embodiments herein has been shown and described above, it is to be understood that above-described embodiment is example
Property, it should not be understood as the limitation to the application, those skilled in the art within the scope of application can be to above-mentioned
Embodiment is changed, modifies, replacement and variant, and scope of the present application is defined by the claims and their equivalents.
Claims (15)
1. a kind of terminal characterized by comprising
Display screen, the display screen include the first viewing area and the second viewing area;And
Imaging device, the imaging device are correspondingly arranged with first viewing area, and the imaging device is for receiving across institute
The optical signal of the first viewing area is stated to be imaged;
When the imaging device is imaged, first viewing area is in OFF state.
2. a kind of terminal characterized by comprising
Display screen, the display screen include the first viewing area and the second viewing area;
Imaging device, the imaging device are correspondingly arranged with first viewing area, and the imaging device is for receiving across institute
The optical signal of the first viewing area is stated to obtain initial pictures;And
Processor, the processor is for handling the initial pictures to obtain final image.
3. terminal according to claim 2, which is characterized in that described when the imaging device obtains the initial pictures
First viewing area is in OFF state, the processor be used to correct the initial pictures by the first algorithm with obtain it is described most
Whole image, first algorithm are related to the screen structure of first viewing area.
4. terminal according to claim 3, which is characterized in that the quantity of first algorithm be it is multiple, each described the
One algorithm is corresponding with an image scene, and the processor is also used to:
Obtain the first algorithm corresponding with the image scene of the initial pictures;And
The initial pictures are corrected by corresponding first algorithm to obtain the final image.
5. terminal according to claim 3, which is characterized in that in calibration, first viewing area is in OFF state,
The imaging device is used to receive optical signal across first viewing area to obtain test image, and described in receiving and not passing through
The optical signal of first viewing area is to obtain reference picture;
The processor is used to obtain first algorithm according to the test image and the reference picture.
6. terminal according to claim 3, which is characterized in that first algorithm includes luminance compensation algorithm, diffraction light
One of spot elimination algorithm, completion algorithm and defogging algorithm are a variety of.
7. terminal according to claim 2, which is characterized in that described when the imaging device obtains the initial pictures
First viewing area shows default picture, the processor be used to correct the initial pictures by the second algorithm with obtain it is described most
Whole image, second algorithm are related to the screen structure of first viewing area and the default picture.
8. according to claim 1 to terminal described in 7 any one, which is characterized in that the display screen includes can be by independence
The the first son screen and the second son screen of control, first viewing area are formed on the first son screen, second viewing area shape
At on the second son screen;Or
The display screen is an independent screen structure, and first viewing area and second viewing area can independently be controlled
System.
9. according to claim 1 to terminal described in 7 any one, which is characterized in that first viewing area is arranged described
The non-corner region of display screen.
10. according to claim 1 to terminal described in 7 any one, which is characterized in that first viewing area is arranged described
The middle position at least one edge of display screen.
11. according to claim 1 to terminal described in 7 any one, which is characterized in that the pixel density of first viewing area
Less than the pixel density of second viewing area.
12. according to claim 1 to terminal described in 7 any one, which is characterized in that the imaging device is not imaged
When, first viewing area and second viewing area are in illuminating state.
13. a kind of control method of terminal, which is characterized in that the terminal includes display screen and imaging device, the display screen
Including the first viewing area and the second viewing area, the imaging device is correspondingly arranged with first viewing area;The control method
Include:
It controls first viewing area and is in OFF state;And
It controls the imaging device and receives the optical signal across first viewing area to be imaged.
14. a kind of image acquiring method, it to be used for terminal, which is characterized in that the terminal includes display screen and imaging device, described
Display screen includes the first viewing area and the second viewing area, and the imaging device is correspondingly arranged with first viewing area, the figure
As acquisition methods include:
It controls first viewing area and is in OFF state;
It controls the imaging device and receives the optical signal across first viewing area to obtain initial pictures;And
The initial pictures are corrected by the first algorithm to obtain final image, first algorithm and first viewing area
It is related to shield structure.
15. a kind of image acquiring method, it to be used for terminal, which is characterized in that the terminal includes display screen and imaging device, described
Display screen includes the first viewing area and the second viewing area, and the imaging device is correspondingly arranged with first viewing area, the figure
As acquisition methods include:
It controls first viewing area and shows default picture;
It controls the imaging device and receives the optical signal across first viewing area to obtain initial pictures;And
The initial pictures are corrected by the second algorithm to obtain final image, second algorithm and first viewing area
Shield structure and the default picture is related.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910556578.8A CN110248004A (en) | 2019-06-25 | 2019-06-25 | Terminal, the control method of terminal and image acquiring method |
PCT/CN2020/096250 WO2020259345A1 (en) | 2019-06-25 | 2020-06-16 | Terminal, control method for same, and image acquisition method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910556578.8A CN110248004A (en) | 2019-06-25 | 2019-06-25 | Terminal, the control method of terminal and image acquiring method |
Publications (1)
Publication Number | Publication Date |
---|---|
CN110248004A true CN110248004A (en) | 2019-09-17 |
Family
ID=67889542
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910556578.8A Pending CN110248004A (en) | 2019-06-25 | 2019-06-25 | Terminal, the control method of terminal and image acquiring method |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN110248004A (en) |
WO (1) | WO2020259345A1 (en) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111383597A (en) * | 2020-03-25 | 2020-07-07 | 武汉华星光电半导体显示技术有限公司 | Pixel circuit and full-screen display equipment |
CN111627369A (en) * | 2020-05-14 | 2020-09-04 | Oppo广东移动通信有限公司 | Display screen display method, display screen assembly and electronic equipment |
WO2020259345A1 (en) * | 2019-06-25 | 2020-12-30 | Oppo广东移动通信有限公司 | Terminal, control method for same, and image acquisition method |
CN112331309A (en) * | 2020-10-29 | 2021-02-05 | 武汉联影医疗科技有限公司 | Medical image film reading device and medical image film reading system |
CN112581874A (en) * | 2019-09-29 | 2021-03-30 | 北京小米移动软件有限公司 | Display panel, display device, display control method and device and terminal equipment |
WO2021258300A1 (en) * | 2020-06-23 | 2021-12-30 | Oppo广东移动通信有限公司 | In-screen photography control method, terminal device and storage medium |
CN114092348A (en) * | 2021-11-18 | 2022-02-25 | 大连海事大学 | Image defogging method and system based on diffraction optical multichannel neural network |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050264689A1 (en) * | 2004-05-26 | 2005-12-01 | Yang Shou A | Display panel having a video camera |
EP2330819A2 (en) * | 2009-11-02 | 2011-06-08 | Sony Corporation | Image display device with imaging unit |
CN102411878A (en) * | 2010-09-21 | 2012-04-11 | 索尼爱立信移动通信日本株式会社 | Sensor-equipped display apparatus and electronic apparatus |
CN106603772A (en) * | 2017-01-26 | 2017-04-26 | 广东欧珀移动通信有限公司 | Electronic device and image acquisition method |
CN107784989A (en) * | 2017-10-27 | 2018-03-09 | 信利光电股份有限公司 | A kind of liquid crystal display device with camera module |
CN108366186A (en) * | 2018-02-09 | 2018-08-03 | 广东欧珀移动通信有限公司 | Electronic device, display screen and camera control method |
CN109451108A (en) * | 2018-11-16 | 2019-03-08 | Oppo广东移动通信有限公司 | Electronic device, the control method of electronic device and control device |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106921767A (en) * | 2017-03-07 | 2017-07-04 | 捷开通讯(深圳)有限公司 | A kind of mobile terminal of screen accounting high |
CN107580092A (en) * | 2017-10-17 | 2018-01-12 | 黄河科技学院 | Comprehensive screen mobile phone and its operating method with camera under transparent screen |
CN108984139A (en) * | 2018-06-22 | 2018-12-11 | 昆山国显光电有限公司 | The control method of display module, terminal and terminal |
CN108900676A (en) * | 2018-08-13 | 2018-11-27 | 华南理工大学 | A kind of comprehensive screen mobile phone and the method for realizing screen display comprehensively |
CN110248004A (en) * | 2019-06-25 | 2019-09-17 | Oppo广东移动通信有限公司 | Terminal, the control method of terminal and image acquiring method |
CN111049973B (en) * | 2019-11-22 | 2021-06-01 | 维沃移动通信有限公司 | Screen display control method, electronic equipment and computer readable storage medium |
-
2019
- 2019-06-25 CN CN201910556578.8A patent/CN110248004A/en active Pending
-
2020
- 2020-06-16 WO PCT/CN2020/096250 patent/WO2020259345A1/en active Application Filing
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050264689A1 (en) * | 2004-05-26 | 2005-12-01 | Yang Shou A | Display panel having a video camera |
EP2330819A2 (en) * | 2009-11-02 | 2011-06-08 | Sony Corporation | Image display device with imaging unit |
CN102411878A (en) * | 2010-09-21 | 2012-04-11 | 索尼爱立信移动通信日本株式会社 | Sensor-equipped display apparatus and electronic apparatus |
CN106603772A (en) * | 2017-01-26 | 2017-04-26 | 广东欧珀移动通信有限公司 | Electronic device and image acquisition method |
CN107784989A (en) * | 2017-10-27 | 2018-03-09 | 信利光电股份有限公司 | A kind of liquid crystal display device with camera module |
CN108366186A (en) * | 2018-02-09 | 2018-08-03 | 广东欧珀移动通信有限公司 | Electronic device, display screen and camera control method |
CN109451108A (en) * | 2018-11-16 | 2019-03-08 | Oppo广东移动通信有限公司 | Electronic device, the control method of electronic device and control device |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2020259345A1 (en) * | 2019-06-25 | 2020-12-30 | Oppo广东移动通信有限公司 | Terminal, control method for same, and image acquisition method |
CN112581874A (en) * | 2019-09-29 | 2021-03-30 | 北京小米移动软件有限公司 | Display panel, display device, display control method and device and terminal equipment |
CN111383597A (en) * | 2020-03-25 | 2020-07-07 | 武汉华星光电半导体显示技术有限公司 | Pixel circuit and full-screen display equipment |
CN111383597B (en) * | 2020-03-25 | 2021-07-06 | 武汉华星光电半导体显示技术有限公司 | Pixel circuit and full-screen display equipment |
CN111627369A (en) * | 2020-05-14 | 2020-09-04 | Oppo广东移动通信有限公司 | Display screen display method, display screen assembly and electronic equipment |
WO2021258300A1 (en) * | 2020-06-23 | 2021-12-30 | Oppo广东移动通信有限公司 | In-screen photography control method, terminal device and storage medium |
CN112331309A (en) * | 2020-10-29 | 2021-02-05 | 武汉联影医疗科技有限公司 | Medical image film reading device and medical image film reading system |
CN114092348A (en) * | 2021-11-18 | 2022-02-25 | 大连海事大学 | Image defogging method and system based on diffraction optical multichannel neural network |
CN114092348B (en) * | 2021-11-18 | 2024-04-05 | 大连海事大学 | Image defogging method and system based on diffractive optical multichannel neural network |
Also Published As
Publication number | Publication date |
---|---|
WO2020259345A1 (en) | 2020-12-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110248004A (en) | Terminal, the control method of terminal and image acquiring method | |
CN112639579B (en) | Spatially resolved dynamic dimming for augmented reality devices | |
US9710887B1 (en) | Display apparatus and method of displaying using context display and projectors | |
US9711114B1 (en) | Display apparatus and method of displaying using projectors | |
US20170208312A1 (en) | Apparatus and method for a dynamic "region of interest" in a display system | |
US9158114B2 (en) | Image display utilizing a variable mask to selectively block image data | |
US20100073544A1 (en) | Image display device, image-capturing device, and program | |
CN106233187A (en) | There is the display device of light modulation panel | |
US20230324744A1 (en) | Geometries for mitigating artifacts in see-through pixel arrays | |
US9549173B2 (en) | Display device | |
CN102760387B (en) | Display floater, display unit and electronic equipment | |
CN109286704A (en) | Electronic equipment | |
CN111077671B (en) | Device control method and device, display device and storage medium | |
JP2017191293A (en) | Peep prevention system | |
CN108345108A (en) | Head-mounted display apparatus, the generation method of three-dimensional image information and device | |
WO2019051228A1 (en) | Camera-based transparent display | |
CN109997067A (en) | Use the display device and method of portable electronic device | |
CN109143575A (en) | Wearable device, anti-peeping display system and anti-peeping display method | |
US20230421883A1 (en) | Reducing imaging stray light crosstalk of camera under lcd screene | |
CN112243117B (en) | Image processing apparatus, method and camera | |
CN205562961U (en) | Image display device and head -mounted apparatus | |
EP3330773B1 (en) | Display apparatus and method of displaying using context display and projectors | |
CN114554050B (en) | Image processing method, device and equipment | |
CN110234004A (en) | Display device, display methods and recording medium | |
CN113888958B (en) | Optical display device, optical display control method and display |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20190917 |
|
RJ01 | Rejection of invention patent application after publication |