US20190312983A1 - Image reading device, image forming apparatus - Google Patents
Image reading device, image forming apparatus Download PDFInfo
- Publication number
- US20190312983A1 US20190312983A1 US16/278,929 US201916278929A US2019312983A1 US 20190312983 A1 US20190312983 A1 US 20190312983A1 US 201916278929 A US201916278929 A US 201916278929A US 2019312983 A1 US2019312983 A1 US 2019312983A1
- Authority
- US
- United States
- Prior art keywords
- pixel data
- light
- pieces
- image
- amount
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 230000002950 deficient Effects 0.000 claims abstract description 17
- 238000012545 processing Methods 0.000 claims description 37
- 230000002159 abnormal effect Effects 0.000 claims description 21
- 238000006243 chemical reaction Methods 0.000 claims description 7
- 230000001678 irradiating effect Effects 0.000 claims description 5
- 238000000034 method Methods 0.000 description 59
- 230000005856 abnormality Effects 0.000 description 40
- 238000012546 transfer Methods 0.000 description 17
- 239000003086 colorant Substances 0.000 description 16
- 239000011521 glass Substances 0.000 description 11
- 238000003860 storage Methods 0.000 description 8
- 238000004140 cleaning Methods 0.000 description 4
- 238000010586 diagram Methods 0.000 description 4
- 230000006870 function Effects 0.000 description 4
- 230000003111 delayed effect Effects 0.000 description 3
- 238000009826 distribution Methods 0.000 description 2
- 238000012423 maintenance Methods 0.000 description 2
- 238000003705 background correction Methods 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000010438 heat treatment Methods 0.000 description 1
- 238000010348 incorporation Methods 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 238000003825 pressing Methods 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N1/00—Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
- H04N1/00002—Diagnosis, testing or measuring; Detecting, analysing or monitoring not otherwise provided for
- H04N1/00007—Diagnosis, testing or measuring; Detecting, analysing or monitoring not otherwise provided for relating to particular apparatus or devices
- H04N1/00013—Reading apparatus
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N1/00—Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
- H04N1/00002—Diagnosis, testing or measuring; Detecting, analysing or monitoring not otherwise provided for
- H04N1/00026—Methods therefor
- H04N1/00037—Detecting, i.e. determining the occurrence of a predetermined state
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N1/00—Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
- H04N1/00002—Diagnosis, testing or measuring; Detecting, analysing or monitoring not otherwise provided for
- H04N1/00026—Methods therefor
- H04N1/00063—Methods therefor using at least a part of the apparatus itself, e.g. self-testing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N1/00—Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
- H04N1/00002—Diagnosis, testing or measuring; Detecting, analysing or monitoring not otherwise provided for
- H04N1/00071—Diagnosis, testing or measuring; Detecting, analysing or monitoring not otherwise provided for characterised by the action taken
- H04N1/00074—Indicating or reporting
Definitions
- the present disclosure relates to an image reading device configured to determine a device state, and relates to an image forming apparatus including the image reading device.
- an image reading device includes a light source and an image sensor, wherein the light source irradiates light to a document sheet or a reference member having a uniform color.
- the reference member is a target of image reading in a shading correction.
- the image sensor includes a plurality of light receiving elements for receiving light reflected on the document sheet or the reference member.
- the plurality of light receiving elements are arranged in alignment along a main scanning direction.
- an image reading device configured to determine whether or not the device state is abnormal. For example, there is known an image reading device configured to average, for each of a plurality of blocks of a light source module, one line of image data obtained from the reference member, and to determine that a block is abnormal when averaged data for each block has a singular value.
- An image reading device includes a light source, an image sensor, a conversion device, an adjustment device, a first representative pixel identifying device, a first determination device, a conspicuous channel identifying device, a second representative pixel identifying device, a second determination device, and a notifying device.
- the light source irradiates light to a document sheet or to a reference portion having a uniform color.
- the image sensor has a plurality of light receiving elements that are arranged in alignment along a main scanning direction, the plurality of light receiving elements being respectively assigned to a plurality of channels and receiving light reflected on the document sheet or the reference portion, the image sensor outputting a plurality of image signals representing amounts of light respectively detected by the plurality of light receiving elements for the plurality of channels.
- the conversion device performs a level adjustment on the plurality of image signals respectively for the plurality of channels based on a level of an adjustment signal that is adjusted in advance, and converts the plurality of image signals after the level adjustment, to a plurality of pieces of pixel data of one line extending in the main scanning direction, the plurality of pieces of pixel data respectively corresponding to the plurality of light receiving elements.
- the adjustment device adjusts the level of the adjustment signal based on a representative value of a plurality of pieces of first pixel data that are a plurality of pieces of pixel data of one line obtained when the light source is turned off.
- the first representative pixel identifying device identifies first representative pixel data that represents a maximum amount of light, among a plurality of pieces of second pixel data that are a plurality of pieces of pixel data of one line obtained when the light source is turned on after the level of the adjustment signal is adjusted.
- the first determination device determines that a device state is normal when each of all values of a plurality of pieces of third pixel data represents an amount of light that is equal to or larger than an amount of light corresponding to a predetermined first threshold, the plurality of pieces of third pixel data being a plurality of pieces of pixel data of one line obtained when the light source is irradiating light to the reference portion after the level of the adjustment signal is adjusted.
- the conspicuous channel identifying device when the first determination device does not determine that the device state is normal, identifies, from among the plurality of channels, one or more conspicuous channels each of which corresponds to a plurality of pieces of third pixel data including a piece of third pixel data representing an amount of light smaller than an amount of light corresponding to the first threshold.
- the second representative pixel identifying device identifies, for each of the one or more conspicuous channels, second representative pixel data that represents a minimum amount of light and third representative pixel data that represents a maximum amount of light, from among the plurality of pieces of third pixel data corresponding to each of the one or more conspicuous channels.
- the second determination device for each of the one or more conspicuous channels, when the third representative pixel data represents an amount of light that is equal to or larger than an amount of light corresponding to at least a second threshold or a third threshold, determines that the device state is defective, and otherwise, determines that the device state is abnormal, the second threshold being determined based on the first representative pixel data, the third threshold being determined for each of the one or more conspicuous channels based on the second representative pixel data and the first representative pixel data.
- the notifying device makes a notification that is different between a case where the device state is determined to be defective and a case where the device state is determined to be abnormal.
- the second threshold corresponds to an amount of light that is larger than an amount of light represented by the first representative pixel data
- the third threshold corresponds to an amount of light that is larger than an amount of light represented by the second representative pixel data.
- An image forming apparatus includes the image reading device and a print processing device.
- the print processing device forms an image on a sheet based on a plurality of pieces of pixel data that are obtained in the image reading device while the light source is irradiating light to the document sheet.
- FIG. 1 is a configuration diagram of an image forming apparatus including an image reading device according to an embodiment of the present disclosure.
- FIG. 2 is a configuration diagram of an image sensor unit and its periphery included in the image reading device according to the embodiment.
- FIG. 3 is a block diagram of a data processing portion included in the image reading device according to the embodiment.
- FIG. 4 is a flowchart showing an example of a procedure of an abnormality determination process executed by the image reading device according to the embodiment.
- FIG. 5 is a diagram showing an example of distribution of line image data in the image reading device according to the embodiment.
- An image reading device 1 constitutes a part of an image forming apparatus 10 .
- the image forming apparatus 10 includes the image reading device 1 and a print processing device 2 . Furthermore, the image forming apparatus 10 includes an operation device 7 a , a display device 7 b , and a data processing device 8 that are common to the image reading device 1 and the print processing device 2 .
- the image forming apparatus 10 is a copier, a printer or a facsimile apparatus having a function of a copier, or a multifunction peripheral having a plurality of image processing functions including an image reading function.
- the image reading device 1 executes a reading process to read an image of a document sheet 9 .
- the print processing device 2 executes a print process to form an image on a sheet.
- a target image of the print process is, for example, an image read by the image reading device 1 , or an image provided by print job data received from a terminal apparatus (not shown).
- the sheet is a sheet-like image formation medium such as a sheet of paper or an OHP sheet.
- the print processing device 2 shown in FIG. 1 forms a toner image on a sheet by an electrophotographic system. It is noted that the print processing device 2 may form an image on a sheet by another system such as an ink jet system.
- the operation device 7 a and the display device 7 b are man-machine interface devices.
- the display device 7 b may include a panel display such as a liquid crystal display
- the operation device 7 a may include a touch panel and operation buttons.
- the data processing device 8 executes data processing on various types of data such as image data obtained via the image reading device 1 . Furthermore, the data processing device 8 controls various types of electric devices of the image processing apparatus 10 based on information input through the operation device 7 a and information detected by various types of sensors.
- the image reading device 1 includes a platen glass 13 , a contact glass 13 a , an image sensor unit 110 , a movable support device 11 , and a platen cover 12 .
- the platen cover 12 incorporates an ADF (Automatic Document Feeder) 14 .
- a document sheet 9 is placed on the platen glass 13 or is conveyed by the ADF 14 .
- the document sheet 9 is a target of image reading.
- the platen cover 12 is supported such that it is displaceable between a position for covering the platen glass 13 and an opening position for opening a space above the platen glass 13 .
- a width direction of the document sheet 9 is referred to as a main scanning direction D 1 .
- a direction in which scanning light moves in the image reading is referred to as a sub scanning direction D 2 .
- the sub scanning direction D 2 is perpendicular to the main scanning direction D 1 . It is noted that FIG. 1 and FIG. 2 show the sub scanning direction D 2 in a case where the document sheet 9 is placed on the platen glass 13 .
- the automatic document sheet conveying device 14 includes a document sheet feeding mechanism 141 and a document sheet conveyance roller 142 .
- the document sheet feeding mechanism 141 feeds a document sheet 9 placed on a document sheet supply tray 121 onto a document sheet conveyance path 140 .
- the document sheet conveyance roller 142 conveys the document sheet 9 along the document sheet conveyance path 140 , and discharges the document sheet 9 onto a document sheet discharge tray 122 .
- the image sensor unit 110 includes a light source 112 , a lens 113 , and an image sensor 114 .
- the light source 112 includes a red light source 112 R, a green light source 112 G, and a blue light source 112 B.
- the image sensor unit 110 of the present embodiment is a CIS module.
- the light source 112 , the lens 113 , and the image sensor 114 are formed to extend along the main scanning direction D 1 .
- the light source 112 emits three colors of light toward a line region that extends along the main scanning direction D 1 on the document sheet 9 .
- the three colors of light are red light, green light, and blue light.
- each of the red light source 112 R, the green light source 112 G, and the blue light source 112 B may be an LED array that includes a plurality of light emitting diodes arranged in alignment along the main scanning direction D 1 .
- the light source 112 emits the three colors of light toward the line region of the document sheet 9 from under the platen glass 13 .
- the three colors of light are irradiated to the document sheet 9 via the platen glass 13 .
- the movable support device 11 moves the image sensor unit 110 along the sub scanning direction D 2 . This allows the light emitted from the light source 112 to scan a surface of the document sheet 9 along the sub scanning direction D 2 .
- the light source 112 emits the three colors of light toward the line region of the document sheet 9 that passes through a reference position P 0 that is in the middle of the document sheet conveyance path 140 .
- the three colors of light are irradiated to the document sheet 9 via the contact glass 13 a.
- the movable support device 11 allows the light emitted from the light source 112 to scan the surface of the document sheet 9 along the sub scanning direction D 2 .
- the ADF 14 allows the light from the light source 112 to scan the surface of the document sheet 9 along the sub scanning direction D 2 .
- the movable support device 11 and the ADF 14 are examples of a scanning device.
- the lens 113 collects light reflected on the line region of the document sheet 9 into a light receiving portion of the image sensor 114 .
- the image sensor 114 is configured to detect an amount of light diffused and reflected on the line region of the document sheet 9 .
- the image sensor 114 is a photoelectric conversion element array including a plurality of light receiving elements aligned along the main scanning direction D 1 .
- each of the light receiving elements is a photoelectric conversion element such as a CMOS image sensor.
- the image sensor 114 outputs an analog line image signal Ia 0 that represents an amount of light diffused and reflected on the line region of the document sheet 9 .
- the line image signal Ia 0 represents the density of a line image that is an image of the line region of the document sheet 9 .
- the image sensor unit 110 When adjustment of the image sensor 114 is made, the image sensor unit 110 is positioned to a position that faces a reference member 15 .
- the reference member 15 is disposed at the reference position P 0 .
- the surface of the reference member 15 has a uniform color.
- the color of the surface of the reference member 15 may be white or pale yellow. It is noted that the surface of the reference member 15 is an example of a reference portion.
- the adjustment of the image sensor 114 is performed in a state where the document sheet 9 is not present at the reference position P 0 .
- the light source 112 emits light to the reference member 15 .
- the image sensor 114 outputs the line image signal Ia 0 that represents an amount of light diffused and reflected on the surface of the reference member 15 .
- the light source 112 irradiates light to the document sheet 9 .
- the light source 112 irradiates light to the reference member 15 .
- another image sensor unit 110 may be fixed at a position that faces a rear surface of the document sheet 9 conveyed along the document sheet conveyance path 140 .
- another reference member 15 is disposed at a position that faces the fixed image sensor unit 110 .
- one of a plurality of document sheet conveyance rollers 142 may serve as the reference member 15 .
- the image sensor 114 is divided into a plurality of channels aligned in the main scanning direction D 1 .
- the image sensor unit 110 is represented by an imaginary line (a two-dot chain line).
- the line image signal Ia 0 output from the image sensor 114 is composed of a plurality of channel image signals Ia that respectively correspond to the plurality of channels.
- the image sensor 114 outputs a plurality of channel image signals Ia( 1 ) to Ia(N) in parallel.
- the image sensor 114 includes a plurality of light receiving elements that receive light reflected on the document sheet 9 or the reference member 15 .
- the plurality of light receiving elements align along the main scanning direction D 1 and are respectively assigned to the plurality of channels.
- the image sensor 114 outputs, in parallel, the plurality of channel image signals Ia( 1 ) to Ia(N) that represent amounts of light respectively detected by the plurality of light receiving elements.
- the channel image signals Ia( 1 ) to Ia(N) are analog.
- the suffixes (1) to (N) of the channel image signals Ia( 1 ) to Ia(N) represent channel numbers, and “N” represents the number of channels. For example, the channel number N exceeds 10 (ten).
- the data processing device 8 executes various types of signal processing on the line image signal Ia 0 output from the image sensor 114 .
- the data processing device 8 includes an AFE (Analog Front End) 81 , a data concatenating portion 82 , and an adjustment portion 83 , wherein the AFE 81 performs predetermined signal processing on the line image signal Ia 0 output from the image sensor 114 .
- AFE Analog Front End
- the AFE 81 , the data concatenating portion 82 , and the adjustment portion 83 are implemented by an electronic circuit(s) such as an ASIC (Application Specific Integrated Circuit).
- the data concatenating portion 82 may be implemented by a processor such as an MPU (Micro Processing Unit) or a DSP (Digital Signal Processor).
- the AFE 81 converts the plurality of channel image signals Ia( 1 ) to Ia(N) respectively to a plurality of pieces of digital channel image data Id( 1 ) to Id(N). During the conversion of the analog signals to the digital data, the AFE 81 performs an offset adjustment on the plurality of channel image signals Ia( 1 ) to Ia(N), and amplifies the signals after the offset adjustment.
- the AFE 81 generates the plurality of pieces of channel image data Id( 1 ) to Id(N) by digitally converting the amplified signals.
- the suffixes (1) to (N) of the plurality of pieces of channel image data Id( 1 ) to Id(N) represent the channel numbers.
- an adjustment signal Sa 0 that has been adjusted in advance is added to the plurality of channel image signals Ia( 1 ) to Ia(N).
- the level of the adjustment signal Sa 0 is adjusted in advance by the adjustment portion 83 that is described below.
- the offset adjustment is an example of a level adjustment that is performed on each of channel image signals Ia based on the level of the adjustment signal Sa 0 .
- the plurality of pieces of channel image data Id( 1 ) to Id(N) are composed of a plurality of pieces of pixel data of one line that align along the main scanning direction D 1 and respectively correspond to the plurality of light receiving elements.
- the plurality of channel image signals Ia( 1 ) to Ia(N) respectively correspond to the plurality of channels.
- the AFE 81 performs the offset adjustment on the plurality of channel image signals Ia( 1 ) to Ia(N) based on the level of the adjustment signal Sa 0 that is adjusted in advance, and converts the signals after the offset adjustment to a plurality of pieces of pixel data.
- the AFE 81 is an example of a conversion device.
- the data concatenating portion 82 generates line image data Id 0 by concatenating a plurality of pieces of channel image data Id. It is noted that the line image data Id 0 is composed of the plurality of pieces of pixel data of one line.
- the adjustment portion 83 adjusts the level of the adjustment signal Sa 0 based on a representative value of the line image data Id 0 that is obtained when the light source 112 is turned off. Specifically, the adjustment portion 83 sets the level of the adjustment signal Sa 0 based on a difference between the representative value of the line image data Id 0 before adjustment and a predetermined reference value.
- the adjustment portion 83 adjusts the level of the adjustment signal Sa 0 so that the minimum value of the line image data Id 0 before adjustment matches the reference value. Subsequently, the adjustment portion 83 supplies the adjustment signal Sa 0 after adjustment to the AFE 81 .
- the image reading device 1 executes a monochrome image reading process, all of the red light source 112 R, the green light source 112 G, and the blue light source 112 B are continuously turned on. This allows white light to be irradiated to the line region of the document sheet 9 or the reference member 15 .
- the red light source 112 R, the green light source 112 G, and the blue light source 112 B are turned on one by one in sequence in a predetermined order. This allows a plurality of colors of light to be irradiated one by one in sequence on the line region of the document sheet 9 or the reference member 15 .
- the data processing device 8 further includes a light emission control portion 86 , a motor control portion 87 , a CPU (Central Processing Unit) 8 a , a RAM (Random Access Memory) 8 b , and a secondary storage device 8 c.
- a light emission control portion 86 a light emission control portion 86 , a motor control portion 87 , a CPU (Central Processing Unit) 8 a , a RAM (Random Access Memory) 8 b , and a secondary storage device 8 c.
- a CPU Central Processing Unit
- RAM Random Access Memory
- the light emission control portion 86 controls turning on and off of the light source 112 .
- the motor control portion 87 controls a first motor 11 m and a second motor 14 m .
- the first motor 11 m is a power source of the movable support device 11 .
- the second motor 14 m is a power source of the ADF 14 . That is, the motor control portion 87 controls the movable support device 11 and the ADF 14 .
- the CPU 8 a controls devices included in the image forming apparatus 10 .
- the CPU 8 a outputs a control signal to each of the AFE 81 , the image sensor 114 , light emission control portion 86 , and the motor control portion 87 so as to start a process thereof.
- the CPU 8 a receives an operation performed by the user on the operation device 7 a , and controls the display device 7 b in accordance with the received operation. For example, the CPU 8 a displays a menu screen or a message on the display device 7 b.
- the secondary storage device 8 c is a computer-readable non-volatile storage device.
- the secondary storage device 8 c stores programs and various types of data. For example, either or both of a flash memory and a hard disk drive are adopted as the secondary storage device 8 c.
- the RAM 8 b is a volatile storage device configured to primarily store: a program executed by the CPU 8 a ; and data that is output and consulted during an execution of the program by the CPU 8 a .
- the RAM 8 b provides data access at a higher speed than the secondary storage device 8 c.
- the print processing device 2 is configured to form an image on a sheet based on the plurality of pieces of pixel data that are obtained when the light source 112 of the image reading device 1 emits light to the document sheet 9 .
- the print processing device 2 includes a sheet conveying portion 3 and an image forming portion 4 .
- the sheet conveying portion 3 feeds a sheet from a sheet storage portion 20 into a sheet conveyance path 30 , and further conveys the sheet along the sheet conveyance path 30 .
- the sheet conveying portion 3 discharges the sheet from the sheet conveyance path 30 onto a discharged tray 22 .
- the image forming portion 4 shown in FIG. 1 forms an image on a sheet by an electrophotographic system.
- the image forming portion 4 includes image creation units 4 x , a laser scanning unit 40 , a transfer device 44 , and a fixing device 46 .
- a drum-like photoconductor 41 rotates, and a charging device 42 charges the surface of the photoconductor 41 uniformly.
- the laser scanning unit 40 writes an electrostatic latent image on the charged surface of the photoconductor 41 .
- the laser scanning unit 40 writes the electrostatic latent image corresponding to the plurality of pieces of pixel data, on the surface of the photoconductor 41 .
- the developing device 43 in each of the image creation units 4 x develops the electrostatic latent image to a toner image.
- the transfer device 44 transfers the toner image from the surface of the photoconductor 41 to the sheet.
- a drum cleaning device 45 in each of the image creation units 4 x removes residual toner from the surface of the photoconductor 41 .
- the fixing device 46 fixes the toner image to the sheet by heating and applying pressure to the toner image on the sheet.
- the print processing device 2 is a color printer including the tandem-type image forming portion 4 . Accordingly, the print processing device 2 includes four image creation units 4 x , and the transfer device 44 includes four primary transfer devices 441 , an intermediate transfer belt 440 , a secondary transfer device 442 , and a belt cleaning device 443 .
- Each of the four image creation units 4 x includes the photoconductor 41 , the charging device 42 , the developing device 43 , and the drum cleaning device 45 .
- the intermediate transfer belt 440 rotates while in contact with surfaces of the four photoconductors 41 .
- the four image creation units 4 x respectively form toner images of colors cyan, magenta, yellow, and black on the surfaces of the photoconductors 41 .
- the four primary transfer devices 441 transfer toner images from the four photoconductors 41 to the intermediate transfer belt 440 . This allows the toner images of the four colors to be overlaid on the intermediate transfer belt 440 such that a color toner image is formed on the intermediate transfer belt 440 .
- the secondary transfer device 442 transfers the toner image from the intermediate transfer belt 440 to the sheet.
- the belt cleaning device 443 removes residual toner from the intermediate transfer belt 440 .
- the image forming portion 4 may be a device adopting another system such as an ink jet system.
- the plurality of light receiving elements are respectively assigned to the plurality of channels.
- the plurality of light receiving elements in the CIS module are respectively assigned to the plurality of channels.
- the image sensor 114 outputs a plurality of image signals that respectively correspond to the plurality of channels.
- An abnormality of the image sensor 114 may occur in a range of channel. However, if it is determined whether or not an abnormality has occurred in the image sensor 114 for each range of channel, it would take time until the result of the determination is obtained. If it takes time until a result of such determination is obtained, start of processing by the image reading device 1 is delayed.
- the device state is distinguished between: an abnormal state where the image sensor 114 or the light source 112 itself does not function normally; and a defective state where, for example, a foreign matter has adhered to the light source 112 .
- the data processing device 8 executes an abnormality determination process that is described below. This allows the image reading device 1 to, in a case where the plurality of light receiving elements of the image sensor 114 are respectively assigned to the plurality of channels, determine, for each channel, whether or not the image reading device 1 is normal, with distinction between the abnormal state and the defective state, while preventing start of image reading from being delayed.
- the data processing device 8 further includes an abnormality determination portion 84 and a main image processing portion 85 .
- the CPU 8 a , the abnormality determination portion 84 , and the light emission control portion 86 execute the abnormality determination process to determine the state of the image sensor 114 and the light source 112 .
- the main image processing portion 85 executes various types of correction processing on the line image data Id 0 , and other image processing.
- the abnormality determination portion 84 and the main image processing portion 85 are implemented by an electronic circuit(s) such as an ASIC.
- the abnormality determination portion 84 and the main image processing portion 85 may be implemented by a processor(s) such as an MPU or a DSP.
- the CPU 8 a executes the abnormality determination process together with the abnormality determination portion 84 and the light emission control portion 86 before the reading process is started.
- the abnormality determination process is started in a state where the image sensor unit 110 is positioned at the reference position P 0 by the movable support device 11 .
- S 1 , S 2 , . . . shown in FIG. 4 are identification signs representing a plurality of steps of the abnormality determination process.
- the vertical axis Vp 0 of the graph in FIG. 5 represents values of the plurality of pieces of pixel data in the line image data Id 0 .
- the CPU 8 a causes the image sensor 114 to execute a dark reading process.
- the image sensor 114 reads an image in a state where the light emission control portion 86 has turned off the light source 112 .
- step S 1 a piece of line image data Id 0 composed of a plurality of pieces of pixel data of one line is obtained in a state where the light source 112 is turned off.
- the plurality of pieces of pixel data of one line constituting a piece of line image data Id 0 obtained in step S 1 are referred to as a plurality of pieces of first pixel data Dp 1 (see FIG. 5 ).
- the adjustment portion 83 adjusts the level of the adjustment signal Sa 0 based on a representative value of the plurality of pieces of first pixel data Dp 1 .
- the adjustment portion 83 executing the process of step S 2 is an example of an adjustment device.
- the adjustment portion 83 adjusts the level of the adjustment signal Sa 0 so that a minimum value Vp 00 of the plurality of pieces of first pixel data Dp 1 before adjustment matches a predetermined reference value Vp 01 (see FIG. 5 ).
- a lookup table or a formula representing a relationship between: differences between the minimum value Vp 00 and the reference value Vp 01 ; and levels of the adjustment signal Sa 0 may be prepared in advance.
- the adjustment portion 83 obtains a level of the adjustment signal Sa 0 by applying a difference between the minimum value Vp 00 and the reference value Vp 01 to the lookup table or the formula, and sets the obtained level of the adjustment signal Sa 0 .
- the CPU 8 a causes the image sensor 114 to execute the dark reading process.
- step S 3 a piece of line image data Id 0 composed of a plurality of pieces of pixel data of one line is obtained in a state where the light source 112 is turned off after adjustment of the level of the adjustment signal Sa 0 .
- the plurality of pieces of pixel data of one line constituting a piece of line image data Id 0 obtained in step S 3 are referred to as a plurality of pieces of second pixel data Dp 2 (see FIG. 5 ).
- the abnormality determination portion 84 identifies, from among the plurality of pieces of second pixel data Dp 2 , first representative pixel data that represents the maximum amount of light.
- VH 1 denotes a value of the first representative pixel data. It is noted that the abnormality determination portion 84 executing the process of step S 4 is an example of a first representative pixel identifying device.
- the light emission control portion 86 turns on the light source 112 .
- the light emission control portion 86 turns on the red light source 112 R, the green light source 112 G, and the blue light source 112 B one by one in sequence.
- the CPU 8 a causes the image sensor 114 to execute a reference reading process.
- the image sensor 114 reads the reference member 15 in a state where the light source 112 is turned on.
- the CPU 8 a each time the red light source 112 R, the green light source 112 G, or the blue light source 112 B is turned on, the CPU 8 a causes the image sensor 114 to execute the reference reading process.
- step S 6 for each of the colors red, green, and blue, a piece of line image data Id 0 composed of a plurality of pieces of pixel data of one line is obtained in a state where the light source 112 is irradiating light to the reference member 15 . Specifically, the line image data Id 0 corresponding to the colors red, green, and blue is obtained.
- the plurality of pieces of pixel data of one line constituting a piece of line image data Id 0 obtained in step S 6 are referred to as a plurality of pieces of third pixel data Dp 3 (see FIG. 5 ).
- the plurality of pieces of third pixel data Dp 3 are obtained after the level of the adjustment signal Sa 0 is adjusted in step S 2 .
- the abnormality determination portion 84 executes a first determination process.
- the abnormality determination portion 84 executing the process of step S 7 is an example of a first determination device. It is noted that the process of step S 7 is executed for each three sets of the plurality of pieces of third pixel data Dp 3 corresponding to the colors red, green, and blue.
- the first determination process it is determined that the device state is normal when each of all values of the plurality of pieces of third pixel data Dp 3 is equal to or larger than a first threshold SL 1 , and otherwise, it is determined that the device state is abnormal.
- the first threshold SL 1 is obtained by adding a predetermined value ⁇ L 0 to the reference value Vp 01 used for the offset adjustment (see FIG. 5 ).
- each of the plurality of pieces of pixel data constituting a piece of line image data Id 0 the larger the value is, the larger the amount of light detected by the image sensor 114 is.
- a value of each of the plurality of pieces of third pixel data Dp 3 is equal to or larger than the first threshold SL 1 , it means that each of the plurality of pieces of third pixel data Dp 3 represents an amount of light that is equal to or larger than an amount of light corresponding to the first threshold SL 1 .
- the first threshold SL 1 having a different value from that of the above-described one is set.
- a value of each of the plurality of pieces of third pixel data Dp 3 is equal to or smaller than the first threshold SL 1 , it means that each of the plurality of pieces of third pixel data Dp 3 represents an amount of light that is equal to or larger than an amount of light corresponding to the first threshold SL 1 .
- the abnormality determination portion 84 Upon determining that the device state is normal, the abnormality determination portion 84 ends the abnormality determination process, and upon determining that the device state is abnormal, the abnormality determination portion 84 moves the process to step S 8 .
- step S 8 the abnormality determination portion 84 executes a process to identify one or more conspicuous channels TC 0 from among the plurality of channels (see FIG. 5 ).
- the abnormality determination portion 84 executing the process of step S 8 is an example of a conspicuous channel identifying device.
- Each conspicuous channel TC 0 among the plurality of channels corresponds to a plurality of pieces of third pixel data Dp 3 including a piece of third pixel data Dp 3 representing an amount of light that is smaller than an amount of light corresponding to the first threshold SL 1 .
- the channel 2 and the channel 5 are conspicuous channels TC 0 .
- step S 8 is executed for each three sets of the plurality of pieces of third pixel data Dp 3 corresponding to the colors red, green, and blue.
- step S 8 is executed when it is determined in step S 7 that the device state is abnormal.
- the abnormality determination portion 84 identifies, for each of the conspicuous channels TC 0 , second representative pixel data that represents the minimum amount of light among the plurality of pieces of third pixel data Dp 3 corresponding to each of the conspicuous channels TC 0 .
- step S 9 is executed for each three sets of the plurality of pieces of third pixel data Dp 3 corresponding to the colors red, green, and blue.
- VL 2 ( 2 ) denotes a value of the second representative pixel data for the channel 2
- VL 2 ( 5 ) denotes a value of the second representative pixel data for the channel 5 .
- the abnormality determination portion 84 identifies, for each of the conspicuous channels TC 0 , third representative pixel data that represents the maximum amount of light among the plurality of pieces of third pixel data Dp 3 corresponding to that conspicuous channel TC 0 .
- step S 10 is executed for each three sets of the plurality of pieces of third pixel data Dp 3 corresponding to the colors red, green, and blue.
- VH 2 ( 2 ) denotes a value of third representative pixel data for the channel 2
- VH 2 ( 5 ) denotes a value of third representative pixel data for the channel 5 .
- abnormality determination portion 84 executing the processes of steps S 9 and S 10 is an example of a second representative pixel identifying device.
- the abnormality determination portion 84 executes a second determination process.
- the abnormality determination portion 84 executing the process of step S 11 is an example of a second determination device.
- step S 11 is executed for each three sets of the plurality of pieces of third pixel data Dp 3 corresponding to the colors red, green, and blue.
- the second threshold SL 2 is set in common to all of the channels.
- the second threshold SL 2 is determined based on the first representative pixel data.
- the second threshold SL 2 corresponds to an amount of light larger than an amount of light represented by the first representative pixel data.
- the second threshold SL 2 is obtained by adding a predetermined value ⁇ L 1 to the value VH 1 of the first representative pixel data (see FIG. 5 ).
- the second threshold SL 2 is calculated by applying the value VH 1 of the first representative pixel data to the following equation (1).
- the third threshold SL 3 ( i ) is set for each of the conspicuous channels TC 0 .
- the third threshold SL 3 ( i ) is determined based on a value VL 2 ( i ) of the second representative pixel data and the value VH 1 of the first representative pixel data.
- the third threshold SL 3 ( i ) corresponds to an amount of light that is larger than an amount of light represented by the second representative pixel data.
- the third threshold SL 3 ( i ) is obtained by adding a predetermined addition value ⁇ L 2 ( i ) that is set for each channel in advance, to the value VL 2 ( i ) of the second representative pixel data (see FIG. 5 ).
- the third threshold SL 3 ( i ) is calculated by applying the value VL 2 ( i ) of the second representative pixel data to the following equation (2).
- the addition value ⁇ L 2 ( i ) is obtained by adding a predetermined adjustment value a 0 to a dark output maximum difference ⁇ V(i) that is set for each of the channels in advance.
- the dark output maximum difference ⁇ V(i) represents the maximum variance of an output value of each of the channels of the image sensor 114 , which is detected when the light source 112 is turned off.
- the dark output maximum difference ⁇ V(i) is a known value included in the specifications of the image sensor 114 .
- the addition value ⁇ L 2 ( i ) is calculated by applying the dark output maximum difference ⁇ V(i) to the following equation (3).
- the dark output maximum difference ⁇ V(i) is a constant that is set for each of the conspicuous channels TC 0
- the adjustment value a 0 is a constant that is common to all the channels.
- the third threshold SL 3 ( i ) that is calculated using the equations (2) and (3) is a value that is obtained by adding a predetermined value that is set for each of the channels, to the second representative pixel data.
- value VH 2 ( i ) of the third representative pixel data when value VH 2 ( i ) of the third representative pixel data is equal to or larger than the second threshold SL 2 , it means that the third representative pixel data represents an amount of light that is equal to or larger than an amount of light corresponding to the second threshold SL 2 .
- value VH 2 ( i ) of the third representative pixel data is equal to or larger than the third threshold SL 3 ( i )
- the third representative pixel data represents an amount of light that is equal to or larger than an amount of light corresponding to the third threshold SL 3 ( i ).
- the device state when the device state is abnormal, it means that the image sensor 114 or the light source 112 itself is not functioning normally in the conspicuous channel TC 0 .
- the device state when the device state is defective, it means that a foreign matter has adhered to a portion of the image sensor 114 or the light source 112 that corresponds to the third representative pixel data. In this case, the device state resumes to a normal state by removing the foreign matter.
- value VH 2 ( 2 ) of the third representative pixel data for the channel 2 is smaller than the second threshold SL 2 and the third threshold SL 3 ( 2 ) of the channel 2 .
- the abnormality determination portion 84 determines that the device state of the channel 2 is abnormal.
- value VH 2 ( 5 ) of the third representative pixel data for the channel 5 is larger than the second threshold SL 2 and the third threshold SL 3 ( 5 ) of the channel 5 .
- the abnormality determination portion 84 determines that the device state of the channel 5 is defective.
- the abnormality determination portion 84 moves the process to step S 12 , and upon determining that the device state is abnormal, the abnormality determination portion 84 moves the process to step S 13 .
- step S 12 is executed on a conspicuous channel(s) TC 0 for which it has been determined that the device state is defective
- step S 13 is executed on a conspicuous channel(s) TC 0 for which it has been determined that the device state is abnormal.
- step S 12 the CPU 8 a makes a first notification to notify that the device state for the conspicuous channel TC 0 is defective. For example, in the first notification, the CPU 8 a displays, on the display device 7 b , a message indicating that a defect has occurred to the device state for the conspicuous channel TC 0 . Thereafter, the CPU 8 a ends the abnormality determination process.
- step S 13 the CPU 8 a makes a second notification to notify that the device state for the conspicuous channel TC 0 is abnormal.
- the CPU 8 a displays, on the display device 7 b , a message indicating that an abnormality has occurred to the device state for the conspicuous channel TC 0 . Thereafter, the CPU 8 a ends the abnormality determination process.
- the CPU 8 a makes a notification that is different between a case where the device state is determined to be defective and a case where the device state is determined to be abnormal. It is noted that the CPU 8 a executing the processes of steps S 12 and S 13 is an example of a notifying device.
- the CPU 8 a When it is determined that the device state is normal, the CPU 8 a , the light emission control portion 86 , and the motor control portion 87 , following the end of the abnormality determination process, cause the light source 112 , the movable support device 11 , and the image sensor 114 to execute the reading process on the document sheet 9 .
- the CPU 8 a when it is determined that the device state is defective or abnormal, the CPU 8 a , the light emission control portion 86 , and the motor control portion 87 do not cause the light source 112 , the movable support device 11 , and the image sensor 114 to execute the reading process on the document sheet 9 .
- the user conveys the content of the first notification or the second notification to a maintenance person. This allows the maintenance person to quickly take appropriate measures to restore the image reading device 1 .
- the first determination process executed in step S 7 is a simple process that does not require data processing for each of the channels. That is, it is possible to determine, in an extremely short time, that the device state is normal.
- the image reading device 1 can prevent start of image reading of the document sheet 9 from being delayed.
- the image reading device 1 is configured to determine the device state for each of the channels by distinguishing between the abnormal state and the defective state (step S 11 ). This contributes to a rapid restoration of the image reading device 1 .
- step S 6 the dark reading process that is to be executed in a state where the light source 112 is turned off, does not need to be executed.
- the abnormality determination portion 84 can also determine, in a relatively short time, that the device state is defective or abnormal.
Abstract
Description
- This application is based upon and claims the benefit of priority from the corresponding Japanese Patent Application No. 2018-074522 filed on Apr. 9, 2018, the entire contents of which are incorporated herein by reference.
- The present disclosure relates to an image reading device configured to determine a device state, and relates to an image forming apparatus including the image reading device.
- In general, an image reading device includes a light source and an image sensor, wherein the light source irradiates light to a document sheet or a reference member having a uniform color. The reference member is a target of image reading in a shading correction.
- The image sensor includes a plurality of light receiving elements for receiving light reflected on the document sheet or the reference member. The plurality of light receiving elements are arranged in alignment along a main scanning direction.
- There is known an image reading device configured to determine whether or not the device state is abnormal. For example, there is known an image reading device configured to average, for each of a plurality of blocks of a light source module, one line of image data obtained from the reference member, and to determine that a block is abnormal when averaged data for each block has a singular value.
- An image reading device according to an aspect of the present disclosure includes a light source, an image sensor, a conversion device, an adjustment device, a first representative pixel identifying device, a first determination device, a conspicuous channel identifying device, a second representative pixel identifying device, a second determination device, and a notifying device. The light source irradiates light to a document sheet or to a reference portion having a uniform color. The image sensor has a plurality of light receiving elements that are arranged in alignment along a main scanning direction, the plurality of light receiving elements being respectively assigned to a plurality of channels and receiving light reflected on the document sheet or the reference portion, the image sensor outputting a plurality of image signals representing amounts of light respectively detected by the plurality of light receiving elements for the plurality of channels. The conversion device performs a level adjustment on the plurality of image signals respectively for the plurality of channels based on a level of an adjustment signal that is adjusted in advance, and converts the plurality of image signals after the level adjustment, to a plurality of pieces of pixel data of one line extending in the main scanning direction, the plurality of pieces of pixel data respectively corresponding to the plurality of light receiving elements. The adjustment device adjusts the level of the adjustment signal based on a representative value of a plurality of pieces of first pixel data that are a plurality of pieces of pixel data of one line obtained when the light source is turned off. The first representative pixel identifying device identifies first representative pixel data that represents a maximum amount of light, among a plurality of pieces of second pixel data that are a plurality of pieces of pixel data of one line obtained when the light source is turned on after the level of the adjustment signal is adjusted. The first determination device determines that a device state is normal when each of all values of a plurality of pieces of third pixel data represents an amount of light that is equal to or larger than an amount of light corresponding to a predetermined first threshold, the plurality of pieces of third pixel data being a plurality of pieces of pixel data of one line obtained when the light source is irradiating light to the reference portion after the level of the adjustment signal is adjusted. The conspicuous channel identifying device, when the first determination device does not determine that the device state is normal, identifies, from among the plurality of channels, one or more conspicuous channels each of which corresponds to a plurality of pieces of third pixel data including a piece of third pixel data representing an amount of light smaller than an amount of light corresponding to the first threshold. The second representative pixel identifying device identifies, for each of the one or more conspicuous channels, second representative pixel data that represents a minimum amount of light and third representative pixel data that represents a maximum amount of light, from among the plurality of pieces of third pixel data corresponding to each of the one or more conspicuous channels. The second determination device, for each of the one or more conspicuous channels, when the third representative pixel data represents an amount of light that is equal to or larger than an amount of light corresponding to at least a second threshold or a third threshold, determines that the device state is defective, and otherwise, determines that the device state is abnormal, the second threshold being determined based on the first representative pixel data, the third threshold being determined for each of the one or more conspicuous channels based on the second representative pixel data and the first representative pixel data. The notifying device makes a notification that is different between a case where the device state is determined to be defective and a case where the device state is determined to be abnormal. The second threshold corresponds to an amount of light that is larger than an amount of light represented by the first representative pixel data, and the third threshold corresponds to an amount of light that is larger than an amount of light represented by the second representative pixel data.
- An image forming apparatus according to an aspect of the present disclosure includes the image reading device and a print processing device. The print processing device forms an image on a sheet based on a plurality of pieces of pixel data that are obtained in the image reading device while the light source is irradiating light to the document sheet.
- This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description with reference where appropriate to the accompanying drawings. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used to limit the scope of the claimed subject matter. Furthermore, the claimed subject matter is not limited to implementations that solve any or all disadvantages noted in any part of this disclosure.
-
FIG. 1 is a configuration diagram of an image forming apparatus including an image reading device according to an embodiment of the present disclosure. -
FIG. 2 is a configuration diagram of an image sensor unit and its periphery included in the image reading device according to the embodiment. -
FIG. 3 is a block diagram of a data processing portion included in the image reading device according to the embodiment. -
FIG. 4 is a flowchart showing an example of a procedure of an abnormality determination process executed by the image reading device according to the embodiment. -
FIG. 5 is a diagram showing an example of distribution of line image data in the image reading device according to the embodiment. - The following describes an embodiment of the present disclosure with reference to the accompanying drawings. It should be noted that the following embodiment is an example of a specific embodiment of the present disclosure and should not limit the technical scope of the present disclosure.
- [Configuration of Image Forming Apparatus 10]
- An
image reading device 1 according to the embodiment of the present disclosure constitutes a part of animage forming apparatus 10. Theimage forming apparatus 10 includes theimage reading device 1 and aprint processing device 2. Furthermore, theimage forming apparatus 10 includes anoperation device 7 a, adisplay device 7 b, and adata processing device 8 that are common to theimage reading device 1 and theprint processing device 2. - For example, the
image forming apparatus 10 is a copier, a printer or a facsimile apparatus having a function of a copier, or a multifunction peripheral having a plurality of image processing functions including an image reading function. - The
image reading device 1 executes a reading process to read an image of adocument sheet 9. Theprint processing device 2 executes a print process to form an image on a sheet. - A target image of the print process is, for example, an image read by the
image reading device 1, or an image provided by print job data received from a terminal apparatus (not shown). The sheet is a sheet-like image formation medium such as a sheet of paper or an OHP sheet. - The
print processing device 2 shown inFIG. 1 forms a toner image on a sheet by an electrophotographic system. It is noted that theprint processing device 2 may form an image on a sheet by another system such as an ink jet system. - The
operation device 7 a and thedisplay device 7 b are man-machine interface devices. For example, thedisplay device 7 b may include a panel display such as a liquid crystal display, and theoperation device 7 a may include a touch panel and operation buttons. - The
data processing device 8 executes data processing on various types of data such as image data obtained via theimage reading device 1. Furthermore, thedata processing device 8 controls various types of electric devices of theimage processing apparatus 10 based on information input through theoperation device 7 a and information detected by various types of sensors. - [Configuration of Image Reading Device 1]
- As shown in
FIG. 1 , theimage reading device 1 includes aplaten glass 13, acontact glass 13 a, animage sensor unit 110, amovable support device 11, and aplaten cover 12. Theplaten cover 12 incorporates an ADF (Automatic Document Feeder) 14. - A
document sheet 9 is placed on theplaten glass 13 or is conveyed by the ADF 14. Thedocument sheet 9 is a target of image reading. Theplaten cover 12 is supported such that it is displaceable between a position for covering theplaten glass 13 and an opening position for opening a space above theplaten glass 13. - In the following description, a width direction of the
document sheet 9 is referred to as a main scanning direction D1. In addition, a direction in which scanning light moves in the image reading is referred to as a sub scanning direction D2. The sub scanning direction D2 is perpendicular to the main scanning direction D1. It is noted thatFIG. 1 andFIG. 2 show the sub scanning direction D2 in a case where thedocument sheet 9 is placed on theplaten glass 13. - The automatic document
sheet conveying device 14 includes a documentsheet feeding mechanism 141 and a documentsheet conveyance roller 142. The documentsheet feeding mechanism 141 feeds adocument sheet 9 placed on a documentsheet supply tray 121 onto a documentsheet conveyance path 140. The documentsheet conveyance roller 142 conveys thedocument sheet 9 along the documentsheet conveyance path 140, and discharges thedocument sheet 9 onto a documentsheet discharge tray 122. - As shown in
FIG. 2 , theimage sensor unit 110 includes alight source 112, alens 113, and animage sensor 114. Thelight source 112 includes ared light source 112R, agreen light source 112G, and a bluelight source 112B. - The
image sensor unit 110 of the present embodiment is a CIS module. Thelight source 112, thelens 113, and theimage sensor 114 are formed to extend along the main scanning direction D1. - The
light source 112 emits three colors of light toward a line region that extends along the main scanning direction D1 on thedocument sheet 9. The three colors of light are red light, green light, and blue light. - For example, each of the
red light source 112R, thegreen light source 112G, and the bluelight source 112B may be an LED array that includes a plurality of light emitting diodes arranged in alignment along the main scanning direction D1. - In a case where the
document sheet 9 is placed on theplaten glass 13, thelight source 112 emits the three colors of light toward the line region of thedocument sheet 9 from under theplaten glass 13. The three colors of light are irradiated to thedocument sheet 9 via theplaten glass 13. In this state, themovable support device 11 moves theimage sensor unit 110 along the sub scanning direction D2. This allows the light emitted from thelight source 112 to scan a surface of thedocument sheet 9 along the sub scanning direction D2. - On the other hand, when the
document sheet 9 is conveyed by theADF 14, thelight source 112 emits the three colors of light toward the line region of thedocument sheet 9 that passes through a reference position P0 that is in the middle of the documentsheet conveyance path 140. In this case, the three colors of light are irradiated to thedocument sheet 9 via thecontact glass 13 a. - That is, when the
document sheet 9 is placed on theplaten glass 13, themovable support device 11 allows the light emitted from thelight source 112 to scan the surface of thedocument sheet 9 along the sub scanning direction D2. In addition, when thedocument sheet 9 is conveyed by theADF 14, theADF 14 allows the light from thelight source 112 to scan the surface of thedocument sheet 9 along the sub scanning direction D2. Themovable support device 11 and theADF 14 are examples of a scanning device. - The
lens 113 collects light reflected on the line region of thedocument sheet 9 into a light receiving portion of theimage sensor 114. Theimage sensor 114 is configured to detect an amount of light diffused and reflected on the line region of thedocument sheet 9. - The
image sensor 114 is a photoelectric conversion element array including a plurality of light receiving elements aligned along the main scanning direction D1. In general, each of the light receiving elements is a photoelectric conversion element such as a CMOS image sensor. - The
image sensor 114 outputs an analog line image signal Ia0 that represents an amount of light diffused and reflected on the line region of thedocument sheet 9. The line image signal Ia0 represents the density of a line image that is an image of the line region of thedocument sheet 9. - When adjustment of the
image sensor 114 is made, theimage sensor unit 110 is positioned to a position that faces areference member 15. In the present embodiment, thereference member 15 is disposed at the reference position P0. - The surface of the
reference member 15 has a uniform color. For example, the color of the surface of thereference member 15 may be white or pale yellow. It is noted that the surface of thereference member 15 is an example of a reference portion. - The adjustment of the
image sensor 114 is performed in a state where thedocument sheet 9 is not present at the reference position P0. In this case, thelight source 112 emits light to thereference member 15. Furthermore, theimage sensor 114 outputs the line image signal Ia0 that represents an amount of light diffused and reflected on the surface of thereference member 15. - That is, when an image of the
document sheet 9 is read, thelight source 112 irradiates light to thedocument sheet 9. On the other hand, when adjustment of theimage sensor 114 is made, thelight source 112 irradiates light to thereference member 15. - It is noted that another
image sensor unit 110 may be fixed at a position that faces a rear surface of thedocument sheet 9 conveyed along the documentsheet conveyance path 140. In this case, anotherreference member 15 is disposed at a position that faces the fixedimage sensor unit 110. In this case, one of a plurality of documentsheet conveyance rollers 142 may serve as thereference member 15. - As shown in
FIG. 3 , theimage sensor 114 is divided into a plurality of channels aligned in the main scanning direction D1. InFIG. 3 , theimage sensor unit 110 is represented by an imaginary line (a two-dot chain line). - The line image signal Ia0 output from the
image sensor 114 is composed of a plurality of channel image signals Ia that respectively correspond to the plurality of channels. Theimage sensor 114 outputs a plurality of channel image signals Ia(1) to Ia(N) in parallel. - That is, the
image sensor 114 includes a plurality of light receiving elements that receive light reflected on thedocument sheet 9 or thereference member 15. The plurality of light receiving elements align along the main scanning direction D1 and are respectively assigned to the plurality of channels. - The
image sensor 114 outputs, in parallel, the plurality of channel image signals Ia(1) to Ia(N) that represent amounts of light respectively detected by the plurality of light receiving elements. The channel image signals Ia(1) to Ia(N) are analog. - In
FIG. 3 , the suffixes (1) to (N) of the channel image signals Ia(1) to Ia(N) represent channel numbers, and “N” represents the number of channels. For example, the channel number N exceeds 10 (ten). - As shown in
FIG. 3 , thedata processing device 8 executes various types of signal processing on the line image signal Ia0 output from theimage sensor 114. For example, thedata processing device 8 includes an AFE (Analog Front End) 81, adata concatenating portion 82, and anadjustment portion 83, wherein theAFE 81 performs predetermined signal processing on the line image signal Ia0 output from theimage sensor 114. - The
AFE 81, thedata concatenating portion 82, and theadjustment portion 83 are implemented by an electronic circuit(s) such as an ASIC (Application Specific Integrated Circuit). In addition, thedata concatenating portion 82 may be implemented by a processor such as an MPU (Micro Processing Unit) or a DSP (Digital Signal Processor). - The
AFE 81 converts the plurality of channel image signals Ia(1) to Ia(N) respectively to a plurality of pieces of digital channel image data Id(1) to Id(N). During the conversion of the analog signals to the digital data, theAFE 81 performs an offset adjustment on the plurality of channel image signals Ia(1) to Ia(N), and amplifies the signals after the offset adjustment. - Furthermore, the
AFE 81 generates the plurality of pieces of channel image data Id(1) to Id(N) by digitally converting the amplified signals. InFIG. 3 , the suffixes (1) to (N) of the plurality of pieces of channel image data Id(1) to Id(N) represent the channel numbers. - In the offset adjustment, an adjustment signal Sa0 that has been adjusted in advance is added to the plurality of channel image signals Ia(1) to Ia(N). The level of the adjustment signal Sa0 is adjusted in advance by the
adjustment portion 83 that is described below. - It is noted that the offset adjustment is an example of a level adjustment that is performed on each of channel image signals Ia based on the level of the adjustment signal Sa0.
- The plurality of pieces of channel image data Id(1) to Id(N) are composed of a plurality of pieces of pixel data of one line that align along the main scanning direction D1 and respectively correspond to the plurality of light receiving elements. The plurality of channel image signals Ia(1) to Ia(N) respectively correspond to the plurality of channels.
- As described above, the
AFE 81 performs the offset adjustment on the plurality of channel image signals Ia(1) to Ia(N) based on the level of the adjustment signal Sa0 that is adjusted in advance, and converts the signals after the offset adjustment to a plurality of pieces of pixel data. TheAFE 81 is an example of a conversion device. - The
data concatenating portion 82 generates line image data Id0 by concatenating a plurality of pieces of channel image data Id. It is noted that the line image data Id0 is composed of the plurality of pieces of pixel data of one line. - The
adjustment portion 83 adjusts the level of the adjustment signal Sa0 based on a representative value of the line image data Id0 that is obtained when thelight source 112 is turned off. Specifically, theadjustment portion 83 sets the level of the adjustment signal Sa0 based on a difference between the representative value of the line image data Id0 before adjustment and a predetermined reference value. - For example, the
adjustment portion 83 adjusts the level of the adjustment signal Sa0 so that the minimum value of the line image data Id0 before adjustment matches the reference value. Subsequently, theadjustment portion 83 supplies the adjustment signal Sa0 after adjustment to theAFE 81. - When the
image reading device 1 executes a monochrome image reading process, all of thered light source 112R, thegreen light source 112G, and the bluelight source 112B are continuously turned on. This allows white light to be irradiated to the line region of thedocument sheet 9 or thereference member 15. - On the other hand, when the
image reading device 1 executes a color image reading process, thered light source 112R, thegreen light source 112G, and the bluelight source 112B are turned on one by one in sequence in a predetermined order. This allows a plurality of colors of light to be irradiated one by one in sequence on the line region of thedocument sheet 9 or thereference member 15. - The
data processing device 8 further includes a lightemission control portion 86, amotor control portion 87, a CPU (Central Processing Unit) 8 a, a RAM (Random Access Memory) 8 b, and asecondary storage device 8 c. - The light
emission control portion 86 controls turning on and off of thelight source 112. Themotor control portion 87 controls afirst motor 11 m and asecond motor 14 m. Thefirst motor 11 m is a power source of themovable support device 11. Thesecond motor 14 m is a power source of theADF 14. That is, themotor control portion 87 controls themovable support device 11 and theADF 14. - The
CPU 8 a controls devices included in theimage forming apparatus 10. For example, theCPU 8 a outputs a control signal to each of theAFE 81, theimage sensor 114, lightemission control portion 86, and themotor control portion 87 so as to start a process thereof. - Furthermore, the
CPU 8 a receives an operation performed by the user on theoperation device 7 a, and controls thedisplay device 7 b in accordance with the received operation. For example, theCPU 8 a displays a menu screen or a message on thedisplay device 7 b. - The
secondary storage device 8 c is a computer-readable non-volatile storage device. Thesecondary storage device 8 c stores programs and various types of data. For example, either or both of a flash memory and a hard disk drive are adopted as thesecondary storage device 8 c. - The
RAM 8 b is a volatile storage device configured to primarily store: a program executed by theCPU 8 a; and data that is output and consulted during an execution of the program by theCPU 8 a. TheRAM 8 b provides data access at a higher speed than thesecondary storage device 8 c. - [Configuration of Print Processing Device 2]
- The
print processing device 2 is configured to form an image on a sheet based on the plurality of pieces of pixel data that are obtained when thelight source 112 of theimage reading device 1 emits light to thedocument sheet 9. - As shown in
FIG. 1 , theprint processing device 2 includes asheet conveying portion 3 and an image forming portion 4. Thesheet conveying portion 3 feeds a sheet from asheet storage portion 20 into asheet conveyance path 30, and further conveys the sheet along thesheet conveyance path 30. In addition, thesheet conveying portion 3 discharges the sheet from thesheet conveyance path 30 onto a dischargedtray 22. - The image forming portion 4 shown in
FIG. 1 forms an image on a sheet by an electrophotographic system. As a result, the image forming portion 4 includesimage creation units 4 x, alaser scanning unit 40, atransfer device 44, and a fixingdevice 46. - In each of the
image creation units 4 x, a drum-like photoconductor 41 rotates, and a chargingdevice 42 charges the surface of thephotoconductor 41 uniformly. - Furthermore, the
laser scanning unit 40 writes an electrostatic latent image on the charged surface of thephotoconductor 41. For example, thelaser scanning unit 40 writes the electrostatic latent image corresponding to the plurality of pieces of pixel data, on the surface of thephotoconductor 41. - The developing
device 43 in each of theimage creation units 4 x develops the electrostatic latent image to a toner image. Thetransfer device 44 transfers the toner image from the surface of thephotoconductor 41 to the sheet. In addition, adrum cleaning device 45 in each of theimage creation units 4 x removes residual toner from the surface of thephotoconductor 41. - The fixing
device 46 fixes the toner image to the sheet by heating and applying pressure to the toner image on the sheet. - As shown in
FIG. 1 , theprint processing device 2 is a color printer including the tandem-type image forming portion 4. Accordingly, theprint processing device 2 includes fourimage creation units 4 x, and thetransfer device 44 includes fourprimary transfer devices 441, an intermediate transfer belt 440, asecondary transfer device 442, and abelt cleaning device 443. - Each of the four
image creation units 4 x includes thephotoconductor 41, the chargingdevice 42, the developingdevice 43, and thedrum cleaning device 45. The intermediate transfer belt 440 rotates while in contact with surfaces of the fourphotoconductors 41. - The four
image creation units 4 x respectively form toner images of colors cyan, magenta, yellow, and black on the surfaces of thephotoconductors 41. The fourprimary transfer devices 441 transfer toner images from the fourphotoconductors 41 to the intermediate transfer belt 440. This allows the toner images of the four colors to be overlaid on the intermediate transfer belt 440 such that a color toner image is formed on the intermediate transfer belt 440. - The
secondary transfer device 442 transfers the toner image from the intermediate transfer belt 440 to the sheet. Thebelt cleaning device 443 removes residual toner from the intermediate transfer belt 440. It is noted that the image forming portion 4 may be a device adopting another system such as an ink jet system. - Meanwhile, in the
image sensor 114, the plurality of light receiving elements are respectively assigned to the plurality of channels. In general, the plurality of light receiving elements in the CIS module are respectively assigned to the plurality of channels. In this case, theimage sensor 114 outputs a plurality of image signals that respectively correspond to the plurality of channels. - An abnormality of the
image sensor 114 may occur in a range of channel. However, if it is determined whether or not an abnormality has occurred in theimage sensor 114 for each range of channel, it would take time until the result of the determination is obtained. If it takes time until a result of such determination is obtained, start of processing by theimage reading device 1 is delayed. - In addition, it is desirable that in determining a device state of the
image reading device 1, the device state is distinguished between: an abnormal state where theimage sensor 114 or thelight source 112 itself does not function normally; and a defective state where, for example, a foreign matter has adhered to thelight source 112. - In the
image reading device 1, thedata processing device 8 executes an abnormality determination process that is described below. This allows theimage reading device 1 to, in a case where the plurality of light receiving elements of theimage sensor 114 are respectively assigned to the plurality of channels, determine, for each channel, whether or not theimage reading device 1 is normal, with distinction between the abnormal state and the defective state, while preventing start of image reading from being delayed. - The
data processing device 8 further includes anabnormality determination portion 84 and a mainimage processing portion 85. TheCPU 8 a, theabnormality determination portion 84, and the lightemission control portion 86 execute the abnormality determination process to determine the state of theimage sensor 114 and thelight source 112. - The main
image processing portion 85 executes various types of correction processing on the line image data Id0, and other image processing. - The
abnormality determination portion 84 and the mainimage processing portion 85 are implemented by an electronic circuit(s) such as an ASIC. In addition, theabnormality determination portion 84 and the mainimage processing portion 85 may be implemented by a processor(s) such as an MPU or a DSP. - For example, when an operation to start the reading process is performed on the
operation device 7 a, theCPU 8 a executes the abnormality determination process together with theabnormality determination portion 84 and the lightemission control portion 86 before the reading process is started. - The abnormality determination process is started in a state where the
image sensor unit 110 is positioned at the reference position P0 by themovable support device 11. - [Abnormality Determination Process]
- Next, an example of a procedure of the abnormality determination process is described with reference to a flowchart shown in
FIG. 4 and a distribution of the line image data Id0 shown inFIG. 5 . - S1, S2, . . . shown in
FIG. 4 are identification signs representing a plurality of steps of the abnormality determination process. In addition, the vertical axis Vp0 of the graph inFIG. 5 represents values of the plurality of pieces of pixel data in the line image data Id0. - <Step S1>
- In the abnormality determination process, first, the
CPU 8 a causes theimage sensor 114 to execute a dark reading process. In the dark reading process, theimage sensor 114 reads an image in a state where the lightemission control portion 86 has turned off thelight source 112. - In step S1, a piece of line image data Id0 composed of a plurality of pieces of pixel data of one line is obtained in a state where the
light source 112 is turned off. - In the following description, the plurality of pieces of pixel data of one line constituting a piece of line image data Id0 obtained in step S1, are referred to as a plurality of pieces of first pixel data Dp1 (see
FIG. 5 ). - <Step S2>
- Subsequently, the
adjustment portion 83 adjusts the level of the adjustment signal Sa0 based on a representative value of the plurality of pieces of first pixel data Dp1. Theadjustment portion 83 executing the process of step S2 is an example of an adjustment device. - As described above, the
adjustment portion 83 adjusts the level of the adjustment signal Sa0 so that a minimum value Vp00 of the plurality of pieces of first pixel data Dp1 before adjustment matches a predetermined reference value Vp01 (seeFIG. 5 ). - For example, a lookup table or a formula representing a relationship between: differences between the minimum value Vp00 and the reference value Vp01; and levels of the adjustment signal Sa0, may be prepared in advance. In this case, the
adjustment portion 83 obtains a level of the adjustment signal Sa0 by applying a difference between the minimum value Vp00 and the reference value Vp01 to the lookup table or the formula, and sets the obtained level of the adjustment signal Sa0. - <Step S3>
- Subsequently, the
CPU 8 a causes theimage sensor 114 to execute the dark reading process. - In step S3, a piece of line image data Id0 composed of a plurality of pieces of pixel data of one line is obtained in a state where the
light source 112 is turned off after adjustment of the level of the adjustment signal Sa0. - In the following description, the plurality of pieces of pixel data of one line constituting a piece of line image data Id0 obtained in step S3, are referred to as a plurality of pieces of second pixel data Dp2 (see
FIG. 5 ). - <Step S4>
- Subsequently, the
abnormality determination portion 84 identifies, from among the plurality of pieces of second pixel data Dp2, first representative pixel data that represents the maximum amount of light. - In
FIG. 5 , VH1 denotes a value of the first representative pixel data. It is noted that theabnormality determination portion 84 executing the process of step S4 is an example of a first representative pixel identifying device. - <Step S5>
- The light
emission control portion 86 turns on thelight source 112. In the present embodiment, the lightemission control portion 86 turns on thered light source 112R, thegreen light source 112G, and the bluelight source 112B one by one in sequence. - <Step S6>
- The
CPU 8 a causes theimage sensor 114 to execute a reference reading process. In the reference reading process, theimage sensor 114 reads thereference member 15 in a state where thelight source 112 is turned on. - In the present embodiment, each time the
red light source 112R, thegreen light source 112G, or the bluelight source 112B is turned on, theCPU 8 a causes theimage sensor 114 to execute the reference reading process. - In step S6, for each of the colors red, green, and blue, a piece of line image data Id0 composed of a plurality of pieces of pixel data of one line is obtained in a state where the
light source 112 is irradiating light to thereference member 15. Specifically, the line image data Id0 corresponding to the colors red, green, and blue is obtained. - In the following description, the plurality of pieces of pixel data of one line constituting a piece of line image data Id0 obtained in step S6, are referred to as a plurality of pieces of third pixel data Dp3 (see
FIG. 5 ). The plurality of pieces of third pixel data Dp3 are obtained after the level of the adjustment signal Sa0 is adjusted in step S2. - <Step S7>
- Subsequently, the
abnormality determination portion 84 executes a first determination process. Theabnormality determination portion 84 executing the process of step S7 is an example of a first determination device. It is noted that the process of step S7 is executed for each three sets of the plurality of pieces of third pixel data Dp3 corresponding to the colors red, green, and blue. - In the first determination process, it is determined that the device state is normal when each of all values of the plurality of pieces of third pixel data Dp3 is equal to or larger than a first threshold SL1, and otherwise, it is determined that the device state is abnormal.
- For example, the first threshold SL1 is obtained by adding a predetermined value ΔL0 to the reference value Vp01 used for the offset adjustment (see
FIG. 5 ). - In the present embodiment, with regard to a value of each of the plurality of pieces of pixel data constituting a piece of line image data Id0, the larger the value is, the larger the amount of light detected by the
image sensor 114 is. As a result, when a value of each of the plurality of pieces of third pixel data Dp3 is equal to or larger than the first threshold SL1, it means that each of the plurality of pieces of third pixel data Dp3 represents an amount of light that is equal to or larger than an amount of light corresponding to the first threshold SL1. - In a case where the smaller the value of each of the plurality of pieces of pixel data is, the larger the amount of light detected by the
image sensor 114 is, the first threshold SL1 having a different value from that of the above-described one is set. In this case, when a value of each of the plurality of pieces of third pixel data Dp3 is equal to or smaller than the first threshold SL1, it means that each of the plurality of pieces of third pixel data Dp3 represents an amount of light that is equal to or larger than an amount of light corresponding to the first threshold SL1. - Upon determining that the device state is normal, the
abnormality determination portion 84 ends the abnormality determination process, and upon determining that the device state is abnormal, theabnormality determination portion 84 moves the process to step S8. - <Step S8>
- In step S8, the
abnormality determination portion 84 executes a process to identify one or more conspicuous channels TC0 from among the plurality of channels (seeFIG. 5 ). Theabnormality determination portion 84 executing the process of step S8 is an example of a conspicuous channel identifying device. - Each conspicuous channel TC0 among the plurality of channels corresponds to a plurality of pieces of third pixel data Dp3 including a piece of third pixel data Dp3 representing an amount of light that is smaller than an amount of light corresponding to the first threshold SL1. In the example shown in
FIG. 5 , thechannel 2 and thechannel 5 are conspicuous channels TC0. - It is noted that the process of step S8 is executed for each three sets of the plurality of pieces of third pixel data Dp3 corresponding to the colors red, green, and blue.
- In the following description, it is assumed that the channel number of the conspicuous channels TC0 is i. It is noted that, as described above, the process of step S8 is executed when it is determined in step S7 that the device state is abnormal.
- <Step S9>
- Subsequently, the
abnormality determination portion 84 identifies, for each of the conspicuous channels TC0, second representative pixel data that represents the minimum amount of light among the plurality of pieces of third pixel data Dp3 corresponding to each of the conspicuous channels TC0. - It is noted that the process of step S9 is executed for each three sets of the plurality of pieces of third pixel data Dp3 corresponding to the colors red, green, and blue.
- In
FIG. 5 , VL2(2) denotes a value of the second representative pixel data for thechannel 2, and VL2(5) denotes a value of the second representative pixel data for thechannel 5. - <Step S10>
- Furthermore, the
abnormality determination portion 84 identifies, for each of the conspicuous channels TC0, third representative pixel data that represents the maximum amount of light among the plurality of pieces of third pixel data Dp3 corresponding to that conspicuous channel TC0. - It is noted that the process of step S10 is executed for each three sets of the plurality of pieces of third pixel data Dp3 corresponding to the colors red, green, and blue.
- In
FIG. 5 , VH2(2) denotes a value of third representative pixel data for thechannel 2, and VH2(5) denotes a value of third representative pixel data for thechannel 5. - It is noted that the
abnormality determination portion 84 executing the processes of steps S9 and S10 is an example of a second representative pixel identifying device. - <Step S11>
- Subsequently, the
abnormality determination portion 84 executes a second determination process. Theabnormality determination portion 84 executing the process of step S11 is an example of a second determination device. - In the second determination process, a determination is made for each of the conspicuous channels TC0. Specifically, when value VH2(i) of the third representative pixel data is equal to or larger than at least a second threshold SL2 or a third threshold SL3(i), it is determined that the device state is defective, and otherwise, it is determined that the device state is abnormal.
- It is noted that the process of step S11 is executed for each three sets of the plurality of pieces of third pixel data Dp3 corresponding to the colors red, green, and blue.
- The second threshold SL2 is set in common to all of the channels. The second threshold SL2 is determined based on the first representative pixel data. The second threshold SL2 corresponds to an amount of light larger than an amount of light represented by the first representative pixel data.
- For example, the second threshold SL2 is obtained by adding a predetermined value ΔL1 to the value VH1 of the first representative pixel data (see
FIG. 5 ). In this case, the second threshold SL2 is calculated by applying the value VH1 of the first representative pixel data to the following equation (1). -
[Math. 1] -
SL2=VH1+ΔL1 (1) - On the other hand, the third threshold SL3(i) is set for each of the conspicuous channels TC0. The third threshold SL3(i) is determined based on a value VL2(i) of the second representative pixel data and the value VH1 of the first representative pixel data.
- For each of the conspicuous channels TC0, the third threshold SL3(i) corresponds to an amount of light that is larger than an amount of light represented by the second representative pixel data.
- For example, the third threshold SL3(i) is obtained by adding a predetermined addition value ΔL2(i) that is set for each channel in advance, to the value VL2(i) of the second representative pixel data (see
FIG. 5 ). In this case, the third threshold SL3(i) is calculated by applying the value VL2(i) of the second representative pixel data to the following equation (2). -
[Math. 2] -
SL3(i)=VL2(i)+ΔL2(i) (2) - The addition value ΔL2(i) is obtained by adding a predetermined adjustment value a0 to a dark output maximum difference ΔV(i) that is set for each of the channels in advance. The dark output maximum difference ΔV(i) represents the maximum variance of an output value of each of the channels of the
image sensor 114, which is detected when thelight source 112 is turned off. The dark output maximum difference ΔV(i) is a known value included in the specifications of theimage sensor 114. In this case, the addition value ΔL2(i) is calculated by applying the dark output maximum difference ΔV(i) to the following equation (3). -
[Math. 3] -
ΔL2(i)=ΔV(i)+α0 (3) - In the equation (3), the dark output maximum difference ΔV(i) is a constant that is set for each of the conspicuous channels TC0, and the adjustment value a0 is a constant that is common to all the channels. Accordingly, the third threshold SL3(i) that is calculated using the equations (2) and (3) is a value that is obtained by adding a predetermined value that is set for each of the channels, to the second representative pixel data.
- In the present embodiment, when value VH2(i) of the third representative pixel data is equal to or larger than the second threshold SL2, it means that the third representative pixel data represents an amount of light that is equal to or larger than an amount of light corresponding to the second threshold SL2.
- Similarly, when value VH2(i) of the third representative pixel data is equal to or larger than the third threshold SL3(i), it means that the third representative pixel data represents an amount of light that is equal to or larger than an amount of light corresponding to the third threshold SL3(i).
- In addition, when the device state is abnormal, it means that the
image sensor 114 or thelight source 112 itself is not functioning normally in the conspicuous channel TC0. - On the other hand, when the device state is defective, it means that a foreign matter has adhered to a portion of the
image sensor 114 or thelight source 112 that corresponds to the third representative pixel data. In this case, the device state resumes to a normal state by removing the foreign matter. - In the example shown in
FIG. 5 , value VH2(2) of the third representative pixel data for thechannel 2 is smaller than the second threshold SL2 and the third threshold SL3(2) of thechannel 2. In this case, theabnormality determination portion 84 determines that the device state of thechannel 2 is abnormal. - In addition, in the example shown in
FIG. 5 , value VH2(5) of the third representative pixel data for thechannel 5 is larger than the second threshold SL2 and the third threshold SL3(5) of thechannel 5. In this case, theabnormality determination portion 84 determines that the device state of thechannel 5 is defective. - Upon determining that the device state is defective, the
abnormality determination portion 84 moves the process to step S12, and upon determining that the device state is abnormal, theabnormality determination portion 84 moves the process to step S13. - It is noted that the result of the determination on the device state may be different for each of the conspicuous channels TC0. In this case, the process of step S12 is executed on a conspicuous channel(s) TC0 for which it has been determined that the device state is defective, and the process of step S13 is executed on a conspicuous channel(s) TC0 for which it has been determined that the device state is abnormal.
- <Step S12>
- In step S12, the
CPU 8 a makes a first notification to notify that the device state for the conspicuous channel TC0 is defective. For example, in the first notification, theCPU 8 a displays, on thedisplay device 7 b, a message indicating that a defect has occurred to the device state for the conspicuous channel TC0. Thereafter, theCPU 8 a ends the abnormality determination process. - <Step S13>
- In step S13, the
CPU 8 a makes a second notification to notify that the device state for the conspicuous channel TC0 is abnormal. For example, in the second notification, theCPU 8 a displays, on thedisplay device 7 b, a message indicating that an abnormality has occurred to the device state for the conspicuous channel TC0. Thereafter, theCPU 8 a ends the abnormality determination process. - That is, the
CPU 8 a makes a notification that is different between a case where the device state is determined to be defective and a case where the device state is determined to be abnormal. It is noted that theCPU 8 a executing the processes of steps S12 and S13 is an example of a notifying device. - When it is determined that the device state is normal, the
CPU 8 a, the lightemission control portion 86, and themotor control portion 87, following the end of the abnormality determination process, cause thelight source 112, themovable support device 11, and theimage sensor 114 to execute the reading process on thedocument sheet 9. - On the other hand, when it is determined that the device state is defective or abnormal, the
CPU 8 a, the lightemission control portion 86, and themotor control portion 87 do not cause thelight source 112, themovable support device 11, and theimage sensor 114 to execute the reading process on thedocument sheet 9. - The user conveys the content of the first notification or the second notification to a maintenance person. This allows the maintenance person to quickly take appropriate measures to restore the
image reading device 1. - As described above, the first determination process executed in step S7 is a simple process that does not require data processing for each of the channels. That is, it is possible to determine, in an extremely short time, that the device state is normal.
- Accordingly, with the above-described configuration, in a case where the plurality of light receiving elements of the
image sensor 114 are respectively assigned to the plurality of channels, theimage reading device 1 can prevent start of image reading of thedocument sheet 9 from being delayed. - Furthermore, the
image reading device 1 is configured to determine the device state for each of the channels by distinguishing between the abnormal state and the defective state (step S11). This contributes to a rapid restoration of theimage reading device 1. - In addition, in the abnormality determination process of the
image reading device 1, after the reference reading process (step S6) is executed in a state where thelight source 112 is turned on, the dark reading process that is to be executed in a state where thelight source 112 is turned off, does not need to be executed. - Accordingly, the
abnormality determination portion 84 can also determine, in a relatively short time, that the device state is defective or abnormal. - It is to be understood that the embodiments herein are illustrative and not restrictive, since the scope of the disclosure is defined by the appended claims rather than by the description preceding them, and all changes that fall within metes and bounds of the claims, or equivalence of such metes and bounds thereof are therefore intended to be embraced by the claims.
Claims (4)
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2018074522A JP6809501B2 (en) | 2018-04-09 | 2018-04-09 | Image reader, image forming device |
JP2018-074522 | 2018-04-09 |
Publications (2)
Publication Number | Publication Date |
---|---|
US10440194B1 US10440194B1 (en) | 2019-10-08 |
US20190312983A1 true US20190312983A1 (en) | 2019-10-10 |
Family
ID=68097482
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/278,929 Expired - Fee Related US10440194B1 (en) | 2018-04-09 | 2019-02-19 | Image reading device, image forming apparatus |
Country Status (2)
Country | Link |
---|---|
US (1) | US10440194B1 (en) |
JP (1) | JP6809501B2 (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2020170115A (en) * | 2019-04-04 | 2020-10-15 | コニカミノルタ株式会社 | Image forming system, control method for image forming system, control program for image forming system |
Family Cites Families (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS59223062A (en) * | 1983-06-01 | 1984-12-14 | Canon Inc | Picture processing device |
US5113251A (en) * | 1989-02-23 | 1992-05-12 | Fuji Xerox Co. | Editing control system and area editing system for image processing equipment |
JP2849627B2 (en) * | 1989-02-27 | 1999-01-20 | 富士ゼロックス株式会社 | Image processing device |
US7128270B2 (en) * | 1999-09-17 | 2006-10-31 | Silverbrook Research Pty Ltd | Scanning device for coded data |
JP4894247B2 (en) | 2005-11-30 | 2012-03-14 | 富士ゼロックス株式会社 | Image reading device |
JP2008124788A (en) * | 2006-11-13 | 2008-05-29 | Brother Ind Ltd | Image reader |
JP5783348B2 (en) * | 2010-12-02 | 2015-09-24 | 富士ゼロックス株式会社 | Control device, control program, and image forming apparatus |
JP5476340B2 (en) * | 2011-06-28 | 2014-04-23 | 京セラドキュメントソリューションズ株式会社 | Image reading apparatus and image reading system |
JP5380563B2 (en) * | 2012-02-24 | 2014-01-08 | 京セラドキュメントソリューションズ株式会社 | Image reading apparatus and image forming apparatus having the same |
JP5690783B2 (en) * | 2012-07-27 | 2015-03-25 | 京セラドキュメントソリューションズ株式会社 | Image reading apparatus and image forming apparatus |
JP5865861B2 (en) * | 2013-03-26 | 2016-02-17 | 京セラドキュメントソリューションズ株式会社 | Image reading apparatus and image forming apparatus |
JP5841972B2 (en) * | 2013-06-28 | 2016-01-13 | 京セラドキュメントソリューションズ株式会社 | Image reading device |
JP6447102B2 (en) * | 2014-12-24 | 2019-01-09 | ブラザー工業株式会社 | Image reading device |
JP6620517B2 (en) * | 2015-10-28 | 2019-12-18 | セイコーエプソン株式会社 | Image reading apparatus and image reading method |
JP6819556B2 (en) * | 2017-11-27 | 2021-01-27 | 京セラドキュメントソリューションズ株式会社 | Image forming device |
-
2018
- 2018-04-09 JP JP2018074522A patent/JP6809501B2/en active Active
-
2019
- 2019-02-19 US US16/278,929 patent/US10440194B1/en not_active Expired - Fee Related
Also Published As
Publication number | Publication date |
---|---|
JP2019186716A (en) | 2019-10-24 |
US10440194B1 (en) | 2019-10-08 |
JP6809501B2 (en) | 2021-01-06 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8810823B2 (en) | Image reading device, an image forming apparatus, and methods for detecting dirt in document reading positions | |
US7817947B2 (en) | Image forming apparatus and correction method of color-misregistration in an image | |
US8179569B2 (en) | Image forming apparatus | |
US11265424B2 (en) | Image forming apparatus | |
US10146162B2 (en) | Image forming apparatus | |
US9383708B1 (en) | Image forming apparatus providing continued processing in the event of sensor failure | |
US10440194B1 (en) | Image reading device, image forming apparatus | |
US8681386B2 (en) | Image reading apparatus and foreign matter detection method | |
US9578194B2 (en) | Image forming apparatus using sync signals for duplex scanning | |
JP6791103B2 (en) | Image reader, image forming device | |
US11575797B2 (en) | Image reading and forming apparatus with streak correction based on image reading mode | |
US10324407B2 (en) | Image forming apparatus | |
JP5225954B2 (en) | Image reading device | |
US11258920B2 (en) | Image diagnostic device, failure diagnostic apparatus, and diagnostic method | |
US9706073B2 (en) | Image reading device and image forming apparatus | |
JP6197566B2 (en) | Image forming apparatus | |
JP5857920B2 (en) | Image forming apparatus | |
US20230161284A1 (en) | Image forming apparatus capable of forming image on sheet with uneven surface, transfer current adjustment method | |
JP2007148079A (en) | Toner density adjusting device and toner density adjusting method | |
JP2009251289A (en) | Image forming apparatus and method of controlling the same | |
JP6459996B2 (en) | Image processing apparatus, image reading apparatus, and foreign object image area detection method | |
JP2017083519A (en) | Image density detecting device and image forming device | |
JP6624096B2 (en) | Image forming apparatus, voltage adjustment method | |
JP6459997B2 (en) | Image processing apparatus, image reading apparatus, and foreign object image area detection method | |
JP2008170578A (en) | Device and method for adjusting toner density |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: KYOCERA DOCUMENT SOLUTIONS INC., JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SHIMATANI, AKIRA;REEL/FRAME:048367/0248 Effective date: 20190130 |
|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
LAPS | Lapse for failure to pay maintenance fees |
Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20231008 |