US20020001407A1 - Method and apparatus for image processing, and storage medium - Google Patents
Method and apparatus for image processing, and storage medium Download PDFInfo
- Publication number
- US20020001407A1 US20020001407A1 US09/010,565 US1056598A US2002001407A1 US 20020001407 A1 US20020001407 A1 US 20020001407A1 US 1056598 A US1056598 A US 1056598A US 2002001407 A1 US2002001407 A1 US 2002001407A1
- Authority
- US
- United States
- Prior art keywords
- image
- reading
- recording medium
- focus information
- recorded
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N1/00—Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
- H04N1/46—Colour picture communication systems
- H04N1/56—Processing of colour picture signals
- H04N1/60—Colour correction or control
- H04N1/6016—Conversion to subtractive colour signals
- H04N1/6022—Generating a fourth subtractive colour signal, e.g. under colour removal, black masking
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N1/00—Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
- H04N1/40—Picture signal circuits
- H04N1/409—Edge or detail enhancement; Noise or error suppression
- H04N1/4092—Edge or detail enhancement
Definitions
- the present invention relates to an image processing apparatus and an image processing method in which image information is read from a recording medium and image processing is suitably performed in accordance with the read image information.
- the invention also pertains to a storage medium in which the above method is stored.
- image information is read from a developed negative or positive film by the scanner and is output to, for example, a printer.
- the above type of film scanner has an image processing section for processing the read image information. Before the image processing section performs image processing on the read image, it is necessary to check the characteristics of the image. This further requires that an image should be first read from a film and then analyzed.
- an image pick-up apparatus which is able to record on the surface of the film various types of information, such as photographic conditions, concerning a photographed picture.
- various types of information such as photographic conditions
- the characteristics of the picture are obtained merely by reading information, such as photographic conditions, recorded on the film without needing to read the picture from the film and extracting its characteristics.
- an image processing apparatus comprising: first reading means for reading an image recorded on a recording medium; second reading means for reading focus information concerning the focus of the image when the image is taken, the image and the focus information being recorded on the recording medium by using different methods; and color-processing means for color-processing the image read by the first reading means in accordance with the focus information read by the second reading means.
- an image processing method comprising: a first reading step of reading an image recorded on a recording medium; a second reading step of reading focus information concerning the focus of the image when the image is taken, the image and the focus information being recorded on the recording medium by using different methods; and a color-processing step of color-processing the image read in the first reading step in accordance with the focus information read in the second reading step.
- a storage medium in which an image processing program is stored in such a manner that the program is readable by a computer, the program comprising: a first reading step of reading an image recorded on a recording medium; a second reading step of reading focus information concerning the focus of the image when the image is taken, the image and the focus information being recorded on the recording medium by using different methods; and a color-processing step of color-processing the image read in the first reading step in accordance with the focus information read in the second reading step.
- an image processing apparatus comprising: first reading means for reading an image recorded on a recording medium; second reading means for reading focus information concerning the focus of the image when the image is taken, the image and the focus information being recorded on the recording medium by using different methods; and filter-processing means for filter-processing the image read by the first reading means in accordance with the focus information read by the second reading means.
- an image processing method comprising: a first reading step of reading an image recorded on a recording medium; a second reading step of reading focus information concerning the focus of the image when the image is taken, the image and the focus information being recorded on the recording medium by using different methods; and a filtering step of filtering the image read in the first reading step in accordance with the focus information read in the second reading step.
- a storage medium in which an image processing program is stored in such a manner that said program is readable by a computer, the program comprising: a first reading step of reading an image recorded on a recording medium; a second reading step of reading focus information concerning the focus of the image when the image is taken, the image and the focus information being recorded on the recording medium by using different methods; and a filtering step of filtering the image read in the first reading step in accordance with the focus information read in the second reading step.
- FIG. 1 is a schematic view illustrating a configuration of an image processing apparatus according to an embodiment of the present invention
- FIG. 2 is a schematic view illustrating a film scanner shown in FIG. 1;
- FIG. 3 is a timing chart illustrating the control timing provided to perform density reproduction in the printer shown in FIG. 1;
- FIG. 4 is a schematic diagram illustrating the flow of an image signal in the image signal processing unit of the image scanner shown in FIG. 1;
- FIG. 5 illustrates the under-color processing section shown in FIG. 4;
- FIG. 6 is a diagram illustrating a technique of obtaining the under-color level from a density histogram
- FIG. 7 is a timing chart illustrating the timing provided for the control signal used in the image signal processing unit shown in FIG. 4;
- FIG. 8 illustrates the relationship between the input data and the output data in the look up table (LUT) shown in FIG. 4;
- FIG. 9 is a flow chart illustrating the process used in the image processing unit shown in FIG. 4.
- FIG. 1 is a schematic sectional view of an image processing apparatus according to a preferred embodiment of the present invention.
- the image processing apparatus has an image scanner 201 and a printer 200 .
- the image scanner 201 reads a document or a film and converts an image signal corresponding to the read image to a digital signal.
- the printer 200 prints in full color on a sheet of paper an image corresponding to the image signal digitized by the scanner 201 .
- the configuration of the image scanner 201 is as follows.
- the document 204 held on the glass 203 by a document press plate 202 is exposed to light from a halogen lamp 205 .
- the light reflected by the document 204 is transmitted to mirrors 206 and 207 , and an image is formed on a three-line sensor (hereinafter referred to as “the CCD”) 210 by a lens unit 208 .
- the lens unit 208 is provided with an infrared-blocking filter 231 .
- the CCD 210 performs color separation on the light reflected from the document 204 and extracts color elements, such as red (R), green (G), and blue (B), of full color information and sends them to an image signal processing unit 209 .
- the sensor arrays of the CCD 210 for reading the color elements are each formed of 5000 pixels. In this manner, the width (297 mm) of an A3-size document, which is the maximum size placable on the document glass 203 , can be read with a 400-dpi resolution.
- the halogen lamp 205 and the mirror 206 are moved at a velocity V, while the mirror 207 is shifted at a velocity 1 ⁇ 2 V, in the directions indicated by the arrows shown in FIG. 1.
- the halogen lamp 205 and the mirrors 206 and 207 are mechanically moved in the perpendicular direction (hereinafter referred to as “the sub-scanning direction”) in relation to the CCD 210 in the electrical scanning direction (hereinafter referred to as “the main scanning direction), thereby scanning the overall surface of the document 204 .
- a standard white plate 211 is used for obtaining corrected data for the read data output from the R, G and B sensors 210 - 1 , 210 - 2 and 210 - 3 of the CCD 210 .
- the standard white plate 211 exhibits substantially uniform reflection characteristics in response to visible light.
- the image data output from the R, G and B sensors 210 - 1 through 210 - 3 is corrected by utilizing the intensity of the light reflected by the white plate 211 .
- FIG. 2 is an exploded view illustrating the internal configuration of the film scanner 227 .
- the film 4 used in this embodiment has an optical recording portion for optically recording a photographic picture and a magnetic recording portion for magnetically recording various photographic conditions (for example, auxiliary information concerning the focal length, the zoom magnification, and the aperture) when the picture is taken.
- a magnetic head 1 serves as reading means for reading the above magnetic recording (photographic data) portion.
- the film scanner 227 includes a hold pad 2 for holding the magnetic head 1 , a film cartridge 3 , a film 4 , a spool 5 for winding the film 4 , and an illuminating device 6 for a scanner 7 .
- the scanner 7 has R, G and B-color line sensors and reads the picture recorded on the film 4 based on the light passing through the film 4 .
- the film 4 is fed to relatively move the scanner 7 in the sub-scanning direction.
- a feeding gear unit 8 is used for feeding the film 4 from the film cartridge 3 .
- a fork 10 is employed for driving the spool 5 to wind the film 4 .
- the film scanner 227 further includes a feeding gear unit 9 and a spool driving gear unit 11 .
- a feeding motor 12 is connected to both the feeding gear unit 9 and the spool driving gear unit 11 .
- the feeding motor 12 is operated to feed the film 4 , and simultaneously, the photographic data magnetically recorded on the film 4 is read by the magnetic head 1 . At this time, the scanner 7 also reads the image (developed image) recorded on the film 4 .
- the image signal processing unit 209 switches between an image signal obtained by reading the document 204 placed on the document glass 203 with the image scanner 201 and an image signal read from the film 4 with the film scanner 227 .
- the selected image signal is then input and electrically processed in the image signal processing unit 209 so as to be separated into the respective color elements, such as magenta (M), cyan (C), yellow (Y), and black (Bk).
- M magenta
- C cyan
- Y yellow
- Bk black
- a semiconductor laser 213 is modulated by a laser driver 212 based on the image signals of the respective colors M, C, Y, and Bk which have been processed and sent from the image signal processing unit 209 .
- Laser light output from the modulated semiconductor laser 213 reaches and scans a photosensitive drum 217 via a polygonal mirror 214 , a f- ⁇ lens 215 , and a mirror 216 .
- an electrostatic latent image corresponding to the input image signals is formed on the photosensitive drum 217 .
- a developer device is formed of a magenta developer unit 219 , a cyan developer unit 220 , a yellow developer unit 221 , and a black developer unit 222 .
- the four developer units 219 through 222 sequentially contact the photosensitive drum 217 to develop the latent image having the respective colors M, C, Y and Bk formed on the photosensitive drum 217 with the corresponding colors of toner.
- a transfer drum 223 takes up and winds a sheet of paper fed from a sheet cassette 224 or 225 so as to transfer the toner image developed on the photosensitive drum 217 to the sheet, thereby forming a color image on the sheet.
- the sheet passes through a fixing unit 226 and is then discharged to the exterior of the image processing apparatus.
- a well-known pulse-width modulation (PWM) method is employed to control the activation time of the semiconductor laser 213 in accordance with the image density signal.
- PWM pulse-width modulation
- the electrostatic latent image having an electric potential in response to the activation time of the semiconductor laser 213 is formed on the photosensitive drum 217 .
- the respective colors of toner by an amount according to the electric potential of the latent image adhere to the photosensitive drum 217 by the developer units 219 through 222 , thereby developing the latent image.
- the developed image is then transferred to a transfer sheet wound around the transfer drum 223 . The image density is thus reproduced.
- FIG. 3 is a timing chart illustrating the control operation for density reproduction performed in the printer 200 .
- 10201 indicates a printer pixel clock, which corresponds to 400-dpi resolution.
- the pixel clock 10201 is generated by the laser driver 212 .
- a 400-line (line/inch) triangular wave 10202 is produced in synchronization with the pixel clock 10201 .
- the triangular wave 10202 and the pixel clock 10201 have the same period.
- 400-dpi resolution image data (digital) signals having the respective M, C, Y and Bk colors, each having a 256-step gradation (8 bits), and a 200-line/400-line switching signal are transmitted to the laser driver 212 from the image signal processing unit 209 in synchronization with the clock of the image data pixels.
- the laser driver 212 synchronizes the above signals with the pixel clock 10201 by using a FIFO memory (not shown).
- the image data having the respective colors, each having 8 bits, is converted into an analog image signal 10203 by a digital-to-analog (D/A) converter (not shown) and is compared with the above-described 400-line triangular wave 10202 , thereby generating a 400-line PWM digital output signal 10204 .
- This digital image data changes from “00H” (H indicates hexadecimal) to “OFFH”, and thus, the 400-line PWM output signal 10204 has a pulse width in accordance with the above values.
- One period of the 400-line PWM output signal 10204 is 63.5 ⁇ m on the photosensitive drum 217 .
- the laser driver 212 produces not only the 400-line triangular wave 10202 but also a 200-line triangular wave 10205 , which has a period twice as long as the 400-line triangular wave 10202 , in synchronization with the pixel clock 10201 .
- a 200-line PWM digital output signal 10206 is generated.
- the 200-line PWM output signal 10206 forms a latent image on the photosensitive drum 207 with a period of 127 ⁇ m, as shown in FIG. 3.
- a 200-line/400-line switching signal 10207 is provided, as shown in FIG. 3.
- the switching signal 10207 is input for each pixel into the laser driver 212 from the image signal processing unit 209 in synchronization with the 400-dpi image data.
- the 200-line/400-line switching signal 10207 is at a logical low level (hereinafter referred to as “the L level”)
- the PWM output signal using the 400-line triangular wave is selected.
- the switching signal 10207 is at a logical high level (hereinafter referred to as “the H level”)
- the PWM output signal using the 200-line triangular wave is selected.
- the configuration of the image signal processing unit 209 of the image scanner 201 is as follows.
- FIG. 4 is a block diagram illustrating the flow of an image signal in the image signal processing unit 9 according to an embodiment of the present invention.
- An image signal output from the CCD 210 is input into an analog signal processing section 101 of the image signal processing unit 209 .
- the gain and the offset of the signal are adjusted.
- the corrected signal is then output to an A/D converter 102 in which the analog signal is converted into digital image signals R 1 , G 1 and B 1 of the respective colors, each having 8 bits.
- the image signals R 1 , G 1 and B 1 are input into a shading correcting section 103 in which the shading of each of the image signals R 1 , G 1 and B 1 is corrected according to a known shading correction method using a signal read from the standard white plate 211 .
- the corrected signals R 2 , G 2 and B 2 are output.
- a clock generator 121 generates a clock (CLK) for each pixel.
- a main-scanning address counter 122 counts the number of clocks (CLK) from the clock generator 121 and generates a pixel address output signal for one line.
- a decoder 123 then decodes the main-scanning address output signal from the address counter 122 and produces the following signals: a CCD drive signal for each line, such as a shift pulse or a reset pulse, a VE signal representing the effective zone of a reading signal for one line output from the CCD 210 , and a line synchronizing signal HSYNC.
- the address counter 122 is reset by the HSYNC signal output from the decoder 123 and starts counting the number of clocks output from the clock generator 121 for a subsequent line.
- An input masking section 106 converts a reading color space, which is determined by the spectral characteristics of the R, G and B filters 210 - 1 , 210 - 2 and 210 - 3 of the CCD 210 , into an NTSC standard color space represented by R 4 , G 4 and B 4 according to the following matrix calculation.
- [ R4 G4 B4 ] [ a11 a12 a13 a21 a22 a23 a31 a32 a33 ] ⁇ [ R3 G3 B3 ]
- An image synthesizing section 1064 respectively synthesizes the image signals R 4 , G 4 and B 4 output from the image scanner 201 with the image signals Rif, Gif and Bif which are output from the film scanner 227 via the input masking section 106 .
- the calculations required for this synthesizing operation are as follows.
- R 40 R 4 ⁇ COMP R+Rif ⁇ ( 1 -COMP R)
- G 40 G 4 ⁇ COMP G+Gif ⁇ ( 1 -COMP G)
- the elements COMP R, COMP G and COMP B are input by the operation performed through an operation section 1101 provided for the image processing apparatus.
- an image read by the image scanner 201 is output.
- a film image read by the film scanner 227 is output. If the values between “0” and “1” are set in COMP R, COMP G and COMP B, an image signal from the image scanner 201 and a film image signal from the film scanner 227 are synthesized and output from the image synthesizing section 1064 .
- An under-color processing section 1065 detects and removes under-color components of the image signals R 40 , G 40 and B 40 output from the image synthesizing section 1064 .
- FIG. 5 is a block diagram illustrating the configuration of the under-color processing section 1065 used in this embodiment.
- the under-color level detector 4001 samples document image signals which are obtained during the prescanning of the image scanner 201 , thereby producing density histograms of the respective image signals R 40 , G 40 and B 40 . Then, according to the respective density histograms, the values of the respective image signals obtained by the following procedure are determined as under-color level signals Rb, Gb and Bb. The density levels higher than a predetermined value “ ⁇ ” and also having a greater frequency than a predetermined ratio are first obtained. Among the above values, the value having the highest density level is then determined.
- FIG. 6 An example of a technique of determining an under-color level signal Rb based on a histogram is shown in FIG. 6.
- the under-color level signals Rb, Gb and Bb obtained by the under-color level detector 4001 are respectively converted into Re, Ge and Be according to the following equations and are input into an under-color removal section 4002 .
- the under-color removal section 4002 the under-color components of the signals Re, Ge and Be are removed by performing a calculation using the following matrix equation, and the resulting signals R 5 , G 5 and B 5 are then output. It should be noted that the signals Re, Ge and Be required for this calculation are input from the under-color level detector 4001 during prescanning.
- [ R5 G5 B5 ] [ 1 0 0 Re 0 1 0 Ge 0 0 1 Be ] ⁇ [ R4 G4 B4 R4 ⁇ G4 ⁇ B4 ]
- the under-color level detector 4001 If the user manually instructs the adjustment of the under-color level through the operation section 1101 , the signal indicating the adjusted level set by the user is input into the under-color level detector 4001 under the control of a CPU 1102 . Then, the under-color level detector 4001 outputs the predetermined values corresponding to the input under-color levels of the respective colors as Re, Ge and Be to the under-color removal section 4002 .
- the above-described detecting operation by the under-color level detector 4001 may be performed using software according to the calculation based on the program of the CPU 1102 .
- a light quantity/density converting section (a LOG converting section) 107 which is formed of a look-up table (LUT) ROM, receives luminance signals R 5 , G 5 and B 5 output from the under-color processing section 1065 and converts them to density signals C 0 , M 0 and Y 0 , which are then output to a line delay memory 108 .
- the line delay memory 108 delays the density signals C 0 , M 0 and Y 0 for a predetermined number of lines.
- image signals C 1 , M 1 and Y 1 and the corresponding UCR signals with respect to the same pixel are simultaneously input into a masking/under color removal (UCR) circuit 109 .
- UCR masking/under color removal
- the masking/UCR circuit 109 extracts a black (K) signal from the three primary color signals C 1 , M 1 and Y 1 and further performs a calculation for correcting apparent contamination of recording color materials (toner) to be used in the printer 200 .
- the resulting signals Y 2 , M 2 , C 2 and K 2 are plane-sequentially output every time the reading operation is performed by the image scanner 201 .
- a main-scanning variable-power circuit 110 enlarges and reduces the image signals Y 2 , M 2 , C 2 and K 2 in the main scanning direction by a known interpolation calculation and outputs the resulting signals Y 3 , M 3 , C 3 and K 3 .
- a spatial-filter processing section (output filter) 111 switches between edge enhancement and smoothing processing in accordance with a two-bit FILTER signal input from a LUT 117 , which is described in detail later, and outputs the resulting signals Y 4 , M 4 , C 4 and K 4 to the printer 200 .
- the image signals Y 4 , M 4 , C 4 and K 4 , and a SEN signal which is a 200-line/400-line switching signal, are transmitted to the laser driver 212 of the printer 200 .
- the signals are then pulse-width modulated (PWM) and are recorded by the printer 200 .
- PWM pulse-width modulated
- FIG. 7 is a timing chart illustrating the control signals used in the image signal processing unit 209 .
- a vertical synchronizing (VSYNC) signal represents the image effective zone in the sub-scanning direction.
- the VSYNC signal indicates logical “1”
- the image reading (scanning) is performed and the output signals (M), (C), (Y) and (K) are sequentially output.
- a VE signal representing the image effective zone in the main scanning direction, provides timing for starting the main scanning operation in the logical “1” zone and is primarily used for controlling the counting of the lines to be delayed.
- a CLOCK signal designates a pixel synchronizing signal used in the image scanner 201 .
- image data read by the CCD 210 is transferred and supplied to the signal processors, such as the A/D converter 102 used in the image signal processing unit 209 .
- the CLOCK signal is also employed for transmitting the image signals and the 200-line/400-line switching (SEN) signal to the laser driver 212 of the printer 200 .
- the subject of the image signal may be estimated as follows. Distant view landscape Middle view portrait Very close view characters
- the focal length information Dis read by the film scanner 227 is input into the LUT 117 .
- Color information COL to be developed in the printer 200 is also input into the LUT 117 , and various signals (UCR signal, FILTER signal, and SEN signal) are generated, as shown in FIG. 8, based on the color information COL and the focal length information Dis.
- FIG. 8 illustrates the relationship between the input data and the output data in the LUT 117 .
- the LUT 117 receives the focal length information Dis (two bits) read by the film scanner 227 .
- the information Dis is defined as follows. Distant view: 3 Middle view: 2 Close view: 1 Very close view: 0
- the developing color COL indicating “1” represents a black color
- the developing color COL designating “0” represents a color other than black.
- the levels of the following types of signals are determined: the masking-coefficient control signal (UCR) to be input into the masking/UCR circuit 109 , the output filter coefficient control signal (FILTER) to be input into the output filter 111 , and the 400-line/200-line switching signal (SEN) input into the printer 200 .
- the resulting signals are then output.
- the UCR signal indicates the intensity level of the UCR operation: the intensity level decreases in the order of 0, 1, 2 and 3.
- the FILTER signal represents the enhancement level of an image edge: the edge enhancement decreases in the order of 0, 1, 2 and 3.
- the SEN signal “0” designates a 400-line signal, while the SEN signal “1” indicates a 200-line signal.
- FIG. 8 shows that the LUT 117 sets the foregoing signals in the following manner when the image processing apparatus of this embodiment is under the standard condition.
- the number of recording lines is set at 200 indicated by the SEN signal “1” unless otherwise instructed by the user through the operation section 1101 .
- the masking/UCR processing circuit 109 generates a black signal K and performs masking according to the UCR control signal (UCR) output from the LUT 117 .
- the equations set out below illustrate masking/UCR equations performed in the masking/UCR processing circuit 109 .
- the minimum value MIN (C, M, Y) of C 1 , M 1 and Y 1 is first obtained to determine a black signal K 1 according to the equation ( 2101 ).
- 4 ⁇ 8 masking is performed by using the equation ( 2102 ) so as to output C 2 , M 2 , Y 2 and K 2 .
- the coefficients m 11 through m 84 indicate masking coefficients determined by the printer 200
- the coefficients k 11 through k 84 represent UCR coefficients determined by the UCR signal.
- the spatial filter processing section (output filter) 111 shown in FIG. 4 two 5-pixel ⁇ 5-pixel spatial filters are provided, and an output signal from one filter is connected to an input signal of the other filter.
- an edge enhancement operation is performed after a smoothing operation. It is thus possible to achieve edge enhancement with a decreased level of Moire fringing and also to output an image of higher quality by a combination of the two types of edge enhancement coefficients.
- FIG. 9 is a flow chart illustrating the operation of generating a Dis signal indicating photographic information and a COL signal representing a developing color, the operation being performed by the CPU 1102 of the image signal processing unit 209 according to an embodiment of the present invention.
- step S 1 the user inputs through the operation section 1101 the information representing a developing color (a color to be printed in the printer 200 ) and the number of lines for performing the PWM operation in the printer 200 . If the input developing color is monochrome, the COL signal is determined to be “1”. On the other hand, if the developing color is other than monochrome, the COL signal is determined to be “0”. The determined COL signal is then output to the LUT 117 in step S 2 .
- step S 3 an image signal is input from the film scanner 227 .
- the image developed on the film 4 set in the film scanner 227 is read with the image scanner 7 , and the photographic data magnetically recorded on the film 4 is further read by the magnetic head 1 .
- step S 4 the level of the Dis signal is determined as discussed above based on the input photographic data, for example, the focal length of the image when the image has been taken.
- the determined Dis signal is output to the LUT 117 in step S 5 .
- the UCR coefficient used in the masking/UCR circuit 109 the type of the spatial filter used in the spatial filter circuit 111 , and the number of lines used for the PWM processing in the printer 200 are determined.
- UCR processing and filtering processing are modified based on photographic data (the focal length in this embodiment) recorded on the film.
- the auxiliary data used in the present invention is not restricted to photographic information, and magnetic information or information in the form of a bar code recorded on part of a document image may be read with the image scanner 201 , and based on the read information, UCR processing or filtering processing may be performed on the document image.
- the photographic data is not limited to the focal length, and various types of information concerning images may be used, or data which is directly usable for masking processing or UCR processing may be employed.
- the information is magnetically stored in a storage medium, the quality of a resulting image is not impaired.
- the auxiliary data may be not only magnetically recorded, but may also be recorded by using a developer which may be only responsive to ultraviolet rays. The advantages offered by the present invention using a developer are similar to those exhibited by the present invention using the magnetically recorded information.
- the present invention may be applied to a system consisting of a plurality of devices (for example, a host computer, an interface, a reader, and a printer) or to a single device (for example, a copying machine or a facsimile machine).
- a plurality of devices for example, a host computer, an interface, a reader, and a printer
- a single device for example, a copying machine or a facsimile machine.
- a storage medium on which a software program code for implementing the function of the foregoing embodiment is recorded may be installed in a system or a device, and the program code stored in the storage medium may be read and run by a computer (or its CPU or MPU) of the system or the device.
- the program code itself read from the storage medium achieves the function of the foregoing embodiment, and accordingly, the storage medium in which the program code is stored may form the present invention.
- the storage mediums for storing the program code may include floppy disks, hard disks, optical disks, magneto-optical disks, CD-ROMs, CD-Rs, magnetic tape, non-volatile memory cards, and ROMs.
- the invention also encompasses the cases where the function of the foregoing embodiment is achieved not only when the program code read from the storage medium is run by a computer, but also when an operating system (OS) run on the computer partially or wholly executes the corresponding processing in response to the instruction of the program code.
- OS operating system
- the invention further encompasses the following case.
- the program code read from a storage medium may be written into a storage device provided in an extended-function board inserted into a computer or an extended-function unit connected to a computer. Then, the CPU provided in the extended-function board or the extended-function unit may partially or wholly perform the corresponding processing in response to the instruction of the program code, thereby implementing the function of the foregoing embodiment.
- An image signal is optically recorded on a recording medium, and at the same time, the corresponding auxiliary information is magnetically recorded on the recording medium. Then, a method for image processing performed on the image signal read from the recording medium is suitably determined based on the auxiliary information. Thus, optimal image processing is performed on the image input from the recording medium.
- an image and the corresponding auxiliary information are recorded on a recording medium by different methods and are read therefrom. Based on the auxiliary information, the image is efficiently color-processed or filtered.
Abstract
An image processor in which image information is read from a recording medium, such as a film used in a camera, with color-processing or filter-processing being efficiently performed in accordance with the contents of the image information. Focus information is also recorded on the recording medium by a method different from the method used for recording the image information. Using the focus information, the image processor suitably switches operation between color processing and filter processing.
Description
- 1. Field of the Invention
- The present invention relates to an image processing apparatus and an image processing method in which image information is read from a recording medium and image processing is suitably performed in accordance with the read image information. The invention also pertains to a storage medium in which the above method is stored.
- 2. Related Background Art
- Hitherto, the following type of film scanner is known: image information is read from a developed negative or positive film by the scanner and is output to, for example, a printer. The above type of film scanner has an image processing section for processing the read image information. Before the image processing section performs image processing on the read image, it is necessary to check the characteristics of the image. This further requires that an image should be first read from a film and then analyzed.
- In order to overcome the above drawback, an image pick-up apparatus is known which is able to record on the surface of the film various types of information, such as photographic conditions, concerning a photographed picture. By using a film in which a picture (image) is recorded by the above type of image pick-up apparatus, the characteristics of the picture are obtained merely by reading information, such as photographic conditions, recorded on the film without needing to read the picture from the film and extracting its characteristics.
- Accordingly, in view of the above background, it is an object of the present invention to efficiently perform color-processing of an image recorded on a recording medium by using focus information concerning the focus of the image when the image is taken, the image and the focus information being recorded on the recording medium by different methods.
- It is another object of the present invention to efficiently perform filtering-processing on an image recorded on a recording medium by using focus information concerning the focus of the image when the image is taken, the image and the focus information being recorded on the recording medium by different methods.
- In order to achieve the above objects, according to one aspect of the present invention, there is provided an image processing apparatus comprising: first reading means for reading an image recorded on a recording medium; second reading means for reading focus information concerning the focus of the image when the image is taken, the image and the focus information being recorded on the recording medium by using different methods; and color-processing means for color-processing the image read by the first reading means in accordance with the focus information read by the second reading means.
- According to another aspect of the present invention, there is provided an image processing method comprising: a first reading step of reading an image recorded on a recording medium; a second reading step of reading focus information concerning the focus of the image when the image is taken, the image and the focus information being recorded on the recording medium by using different methods; and a color-processing step of color-processing the image read in the first reading step in accordance with the focus information read in the second reading step.
- According to still another aspect of the present invention, there is provided a storage medium in which an image processing program is stored in such a manner that the program is readable by a computer, the program comprising: a first reading step of reading an image recorded on a recording medium; a second reading step of reading focus information concerning the focus of the image when the image is taken, the image and the focus information being recorded on the recording medium by using different methods; and a color-processing step of color-processing the image read in the first reading step in accordance with the focus information read in the second reading step.
- According to a further aspect of the present invention, there is provided an image processing apparatus comprising: first reading means for reading an image recorded on a recording medium; second reading means for reading focus information concerning the focus of the image when the image is taken, the image and the focus information being recorded on the recording medium by using different methods; and filter-processing means for filter-processing the image read by the first reading means in accordance with the focus information read by the second reading means.
- According to a further aspect of the present invention, there is provided an image processing method comprising: a first reading step of reading an image recorded on a recording medium; a second reading step of reading focus information concerning the focus of the image when the image is taken, the image and the focus information being recorded on the recording medium by using different methods; and a filtering step of filtering the image read in the first reading step in accordance with the focus information read in the second reading step.
- According to a yet further aspect of the present invention, there is provided a storage medium in which an image processing program is stored in such a manner that said program is readable by a computer, the program comprising: a first reading step of reading an image recorded on a recording medium; a second reading step of reading focus information concerning the focus of the image when the image is taken, the image and the focus information being recorded on the recording medium by using different methods; and a filtering step of filtering the image read in the first reading step in accordance with the focus information read in the second reading step.
- Further objects, features and advantages of the present invention will become apparent from the following description of the preferred embodiments with reference to the attached drawings.
- FIG. 1 is a schematic view illustrating a configuration of an image processing apparatus according to an embodiment of the present invention;
- FIG. 2 is a schematic view illustrating a film scanner shown in FIG. 1;
- FIG. 3 is a timing chart illustrating the control timing provided to perform density reproduction in the printer shown in FIG. 1;
- FIG. 4 is a schematic diagram illustrating the flow of an image signal in the image signal processing unit of the image scanner shown in FIG. 1;
- FIG. 5 illustrates the under-color processing section shown in FIG. 4;
- FIG. 6 is a diagram illustrating a technique of obtaining the under-color level from a density histogram;
- FIG. 7 is a timing chart illustrating the timing provided for the control signal used in the image signal processing unit shown in FIG. 4;
- FIG. 8 illustrates the relationship between the input data and the output data in the look up table (LUT) shown in FIG. 4; and
- FIG. 9 is a flow chart illustrating the process used in the image processing unit shown in FIG. 4.
- A preferred embodiment of the present invention is described in detail below with reference to the accompanying drawings.
- FIG. 1 is a schematic sectional view of an image processing apparatus according to a preferred embodiment of the present invention. In FIG. 1, the image processing apparatus has an
image scanner 201 and aprinter 200. Theimage scanner 201 reads a document or a film and converts an image signal corresponding to the read image to a digital signal. Theprinter 200 prints in full color on a sheet of paper an image corresponding to the image signal digitized by thescanner 201. - The configuration of the
image scanner 201 is as follows. - In order to read an image of an original document204 placed on an original-document glass (platen) 203, the document 204 held on the
glass 203 by adocument press plate 202 is exposed to light from ahalogen lamp 205. The light reflected by the document 204 is transmitted tomirrors blocking filter 231. - The
CCD 210 performs color separation on the light reflected from the document 204 and extracts color elements, such as red (R), green (G), and blue (B), of full color information and sends them to an imagesignal processing unit 209. The sensor arrays of theCCD 210 for reading the color elements are each formed of 5000 pixels. In this manner, the width (297 mm) of an A3-size document, which is the maximum size placable on thedocument glass 203, can be read with a 400-dpi resolution. - The
halogen lamp 205 and themirror 206 are moved at a velocity V, while themirror 207 is shifted at a velocity ½ V, in the directions indicated by the arrows shown in FIG. 1. In this fashion, thehalogen lamp 205 and themirrors CCD 210 in the electrical scanning direction (hereinafter referred to as “the main scanning direction), thereby scanning the overall surface of the document 204. - A standard
white plate 211 is used for obtaining corrected data for the read data output from the R, G and B sensors 210-1, 210-2 and 210-3 of theCCD 210. The standardwhite plate 211 exhibits substantially uniform reflection characteristics in response to visible light. In this embodiment, the image data output from the R, G and B sensors 210-1 through 210-3 is corrected by utilizing the intensity of the light reflected by thewhite plate 211. - The operation of reading an image signal and magnetic information from the
film 4 with afilm scanner 227 is described below. - FIG. 2 is an exploded view illustrating the internal configuration of the
film scanner 227. Thefilm 4 used in this embodiment has an optical recording portion for optically recording a photographic picture and a magnetic recording portion for magnetically recording various photographic conditions (for example, auxiliary information concerning the focal length, the zoom magnification, and the aperture) when the picture is taken. - Referring to FIG. 2, a
magnetic head 1 serves as reading means for reading the above magnetic recording (photographic data) portion. Thefilm scanner 227 includes ahold pad 2 for holding themagnetic head 1, afilm cartridge 3, afilm 4, aspool 5 for winding thefilm 4, and anilluminating device 6 for ascanner 7. Thescanner 7 has R, G and B-color line sensors and reads the picture recorded on thefilm 4 based on the light passing through thefilm 4. Thefilm 4 is fed to relatively move thescanner 7 in the sub-scanning direction. Afeeding gear unit 8 is used for feeding thefilm 4 from thefilm cartridge 3. Afork 10 is employed for driving thespool 5 to wind thefilm 4. Thefilm scanner 227 further includes afeeding gear unit 9 and a spooldriving gear unit 11. Afeeding motor 12 is connected to both thefeeding gear unit 9 and the spooldriving gear unit 11. - In the
film scanner 227 constructed as described above, thefeeding motor 12 is operated to feed thefilm 4, and simultaneously, the photographic data magnetically recorded on thefilm 4 is read by themagnetic head 1. At this time, thescanner 7 also reads the image (developed image) recorded on thefilm 4. - A description is given below of the image
signal processing unit 209 for receiving the image signal from theimage scanner 201 or thefilm scanner 227. - Referring back to FIG. 1, the image
signal processing unit 209 switches between an image signal obtained by reading the document 204 placed on thedocument glass 203 with theimage scanner 201 and an image signal read from thefilm 4 with thefilm scanner 227. The selected image signal is then input and electrically processed in the imagesignal processing unit 209 so as to be separated into the respective color elements, such as magenta (M), cyan (C), yellow (Y), and black (Bk). The separated color elements are then output to theprinter 200. When an image is printed from theimage scanner 201, data concerning one of the color elements, M, C, Y and Bk, is output to theprinter 200 by scanning a document one time (plane-sequential image formation), and the entire image contained in one sheet is printed in full color by scanning the document four times. - In the
printer 200, asemiconductor laser 213 is modulated by alaser driver 212 based on the image signals of the respective colors M, C, Y, and Bk which have been processed and sent from the imagesignal processing unit 209. Laser light output from the modulatedsemiconductor laser 213 reaches and scans aphotosensitive drum 217 via apolygonal mirror 214, a f-θ lens 215, and amirror 216. Thus, an electrostatic latent image corresponding to the input image signals is formed on thephotosensitive drum 217. - A developer device is formed of a
magenta developer unit 219, acyan developer unit 220, ayellow developer unit 221, and ablack developer unit 222. The fourdeveloper units 219 through 222 sequentially contact thephotosensitive drum 217 to develop the latent image having the respective colors M, C, Y and Bk formed on thephotosensitive drum 217 with the corresponding colors of toner. Atransfer drum 223 takes up and winds a sheet of paper fed from asheet cassette photosensitive drum 217 to the sheet, thereby forming a color image on the sheet. - After the toner images having the respective four colors M, C, Y and Bk are sequentially transferred to the sheet, the sheet passes through a fixing
unit 226 and is then discharged to the exterior of the image processing apparatus. - An explanation is given below of a density reproducing method used in the
printer 200 of the image processing apparatus according to an embodiment of the present invention. - In the
printer 200 of this embodiment, in order to precisely reproduce image density, a well-known pulse-width modulation (PWM) method is employed to control the activation time of thesemiconductor laser 213 in accordance with the image density signal. In this method, the electrostatic latent image having an electric potential in response to the activation time of thesemiconductor laser 213 is formed on thephotosensitive drum 217. Then, the respective colors of toner by an amount according to the electric potential of the latent image adhere to thephotosensitive drum 217 by thedeveloper units 219 through 222, thereby developing the latent image. The developed image is then transferred to a transfer sheet wound around thetransfer drum 223. The image density is thus reproduced. - FIG. 3 is a timing chart illustrating the control operation for density reproduction performed in the
printer 200. - In FIG. 3, 10201 indicates a printer pixel clock, which corresponds to 400-dpi resolution. The
pixel clock 10201 is generated by thelaser driver 212. Moreover, a 400-line (line/inch)triangular wave 10202 is produced in synchronization with thepixel clock 10201. Thetriangular wave 10202 and thepixel clock 10201 have the same period. - 400-dpi resolution image data (digital) signals having the respective M, C, Y and Bk colors, each having a 256-step gradation (8 bits), and a 200-line/400-line switching signal are transmitted to the
laser driver 212 from the imagesignal processing unit 209 in synchronization with the clock of the image data pixels. Thelaser driver 212 synchronizes the above signals with thepixel clock 10201 by using a FIFO memory (not shown). The image data having the respective colors, each having 8 bits, is converted into ananalog image signal 10203 by a digital-to-analog (D/A) converter (not shown) and is compared with the above-described 400-linetriangular wave 10202, thereby generating a 400-line PWMdigital output signal 10204. This digital image data changes from “00H” (H indicates hexadecimal) to “OFFH”, and thus, the 400-linePWM output signal 10204 has a pulse width in accordance with the above values. One period of the 400-linePWM output signal 10204 is 63.5 μm on thephotosensitive drum 217. - Further, the
laser driver 212 produces not only the 400-linetriangular wave 10202 but also a 200-linetriangular wave 10205, which has a period twice as long as the 400-linetriangular wave 10202, in synchronization with thepixel clock 10201. Upon comparison of the 200-linetriangular wave 10205 with the 400-dpianalog image signal 10203, a 200-line PWMdigital output signal 10206 is generated. The 200-linePWM output signal 10206 forms a latent image on thephotosensitive drum 207 with a period of 127 μm, as shown in FIG. 3. - Upon comparing density reproduction used by the 200-line triangular wave with density reproduction used by the 400-line triangular wave, a higher level of gradation reproducibility is achieved by the 200-line triangular wave because the minimum unit of the former wave for density reproduction is 127 μm, which is twice as long as the latter wave. In terms of resolution, however, the 400-line triangular wave that reproduces density by a 63.5 μm unit is more suitable for higher-resolution image recording. Consequently, 200-line PWM recording achieves a higher level of gradation reproduction, while 400-line PWM recording exhibits a higher level of resolution. To perform optimal density reproduction, therefore, PWM recording is switched between the 200-line triangular wave and the 400-line triangular wave in accordance with the nature of an image to be recorded.
- For performing the above PWM switching operation, a 200-line/400-
line switching signal 10207 is provided, as shown in FIG. 3. Theswitching signal 10207 is input for each pixel into thelaser driver 212 from the imagesignal processing unit 209 in synchronization with the 400-dpi image data. When the 200-line/400-line switching signal 10207 is at a logical low level (hereinafter referred to as “the L level”), the PWM output signal using the 400-line triangular wave is selected. In contrast, when theswitching signal 10207 is at a logical high level (hereinafter referred to as “the H level”), the PWM output signal using the 200-line triangular wave is selected. - The configuration of the image
signal processing unit 209 of theimage scanner 201 is as follows. - FIG. 4 is a block diagram illustrating the flow of an image signal in the image
signal processing unit 9 according to an embodiment of the present invention. An image signal output from theCCD 210 is input into an analogsignal processing section 101 of the imagesignal processing unit 209. In the analogsignal processing section 101, the gain and the offset of the signal are adjusted. The corrected signal is then output to an A/D converter 102 in which the analog signal is converted into digital image signals R1, G1 and B1 of the respective colors, each having 8 bits. Subsequently, the image signals R1, G1 and B1 are input into ashading correcting section 103 in which the shading of each of the image signals R1, G1 and B1 is corrected according to a known shading correction method using a signal read from the standardwhite plate 211. The corrected signals R2, G2 and B2 are output. - Meanwhile, a
clock generator 121 generates a clock (CLK) for each pixel. A main-scanningaddress counter 122 counts the number of clocks (CLK) from theclock generator 121 and generates a pixel address output signal for one line. Adecoder 123 then decodes the main-scanning address output signal from theaddress counter 122 and produces the following signals: a CCD drive signal for each line, such as a shift pulse or a reset pulse, a VE signal representing the effective zone of a reading signal for one line output from theCCD 210, and a line synchronizing signal HSYNC. Theaddress counter 122 is reset by the HSYNC signal output from thedecoder 123 and starts counting the number of clocks output from theclock generator 121 for a subsequent line. -
- An
image synthesizing section 1064 respectively synthesizes the image signals R4, G4 and B4 output from theimage scanner 201 with the image signals Rif, Gif and Bif which are output from thefilm scanner 227 via theinput masking section 106. The calculations required for this synthesizing operation are as follows. - R40=R4×COMP R+Rif×(1-COMP R)
- G40=G4×COMP G+Gif×(1-COMP G)
- B40=B4×COMP B+Bif×(1-COMP B)
- The elements COMP R, COMP G and COMP B are input by the operation performed through an
operation section 1101 provided for the image processing apparatus. When all of the elements COMP R, COMP G and COMP B indicate “1”, an image read by theimage scanner 201 is output. In contrast, when all of the elements COMP R, COMP G and COMP B indicate “0”, a film image read by thefilm scanner 227 is output. If the values between “0” and “1” are set in COMP R, COMP G and COMP B, an image signal from theimage scanner 201 and a film image signal from thefilm scanner 227 are synthesized and output from theimage synthesizing section 1064. - An under-
color processing section 1065 detects and removes under-color components of the image signals R40, G40 and B40 output from theimage synthesizing section 1064. - FIG. 5 is a block diagram illustrating the configuration of the under-
color processing section 1065 used in this embodiment. - In FIG. 5, when the
operation section 1101 sets an AE mode signal representing the removal of the under-color components, the under-color level detector 4001 samples document image signals which are obtained during the prescanning of theimage scanner 201, thereby producing density histograms of the respective image signals R40, G40 and B40. Then, according to the respective density histograms, the values of the respective image signals obtained by the following procedure are determined as under-color level signals Rb, Gb and Bb. The density levels higher than a predetermined value “α” and also having a greater frequency than a predetermined ratio are first obtained. Among the above values, the value having the highest density level is then determined. - An example of a technique of determining an under-color level signal Rb based on a histogram is shown in FIG. 6.
- The under-color level signals Rb, Gb and Bb obtained by the under-
color level detector 4001 are respectively converted into Re, Ge and Be according to the following equations and are input into an under-color removal section 4002. - Re=(255−Rb)×255×255/(Rb×Gb×Bb)
- Ge=(255−Gb)×255×255/(Rb×Gb×Bb)
- Be=(255−Bb)×255×255/(Rb×Gb×Bb)
- In the under-
color removal section 4002, the under-color components of the signals Re, Ge and Be are removed by performing a calculation using the following matrix equation, and the resulting signals R5, G5 and B5 are then output. It should be noted that the signals Re, Ge and Be required for this calculation are input from the under-color level detector 4001 during prescanning. - If the user manually instructs the adjustment of the under-color level through the
operation section 1101, the signal indicating the adjusted level set by the user is input into the under-color level detector 4001 under the control of aCPU 1102. Then, the under-color level detector 4001 outputs the predetermined values corresponding to the input under-color levels of the respective colors as Re, Ge and Be to the under-color removal section 4002. - The above-described detecting operation by the under-
color level detector 4001 may be performed using software according to the calculation based on the program of theCPU 1102. - Referring back to FIG. 4, a light quantity/density converting section (a LOG converting section)107, which is formed of a look-up table (LUT) ROM, receives luminance signals R5, G5 and B5 output from the under-
color processing section 1065 and converts them to density signals C0, M0 and Y0, which are then output to aline delay memory 108. Theline delay memory 108 delays the density signals C0, M0 and Y0 for a predetermined number of lines. As a consequence, image signals C1, M1 and Y1 and the corresponding UCR signals with respect to the same pixel are simultaneously input into a masking/under color removal (UCR)circuit 109. - The masking/
UCR circuit 109 extracts a black (K) signal from the three primary color signals C1, M1 and Y1 and further performs a calculation for correcting apparent contamination of recording color materials (toner) to be used in theprinter 200. The resulting signals Y2, M2, C2 and K2, each having 8 bits, are plane-sequentially output every time the reading operation is performed by theimage scanner 201. - A main-scanning variable-
power circuit 110 enlarges and reduces the image signals Y2, M2, C2 and K2 in the main scanning direction by a known interpolation calculation and outputs the resulting signals Y3, M3, C3 and K3. Further, a spatial-filter processing section (output filter) 111 switches between edge enhancement and smoothing processing in accordance with a two-bit FILTER signal input from aLUT 117, which is described in detail later, and outputs the resulting signals Y4, M4, C4 and K4 to theprinter 200. - The image signals Y4, M4, C4 and K4, and a SEN signal, which is a 200-line/400-line switching signal, are transmitted to the
laser driver 212 of theprinter 200. The signals are then pulse-width modulated (PWM) and are recorded by theprinter 200. - FIG. 7 is a timing chart illustrating the control signals used in the image
signal processing unit 209. - In FIG. 7, a vertical synchronizing (VSYNC) signal represents the image effective zone in the sub-scanning direction. When the VSYNC signal indicates logical “1”, the image reading (scanning) is performed and the output signals (M), (C), (Y) and (K) are sequentially output. A VE signal, representing the image effective zone in the main scanning direction, provides timing for starting the main scanning operation in the logical “1” zone and is primarily used for controlling the counting of the lines to be delayed. A CLOCK signal designates a pixel synchronizing signal used in the
image scanner 201. When the CLOCK signal is at a rising edge timing “0”→“1”, image data read by theCCD 210 is transferred and supplied to the signal processors, such as the A/D converter 102 used in the imagesignal processing unit 209. The CLOCK signal is also employed for transmitting the image signals and the 200-line/400-line switching (SEN) signal to thelaser driver 212 of theprinter 200. - A description is now given of the image processing executed on photographic data transmitted from the
film scanner 227 according to an embodiment of the present invention. - When an image recorded on the film is read and input with the
film scanner 227, photographic data concerning the input image signal is simultaneously read by themagnetic head 1. In this embodiment, as an example of the photographic data, the focal length determined during a photographic operation is described. - Based on the photographic data concerning the focal length, the subject of the image signal may be estimated as follows.
Distant view landscape Middle view portrait Very close view characters - As a closer view is focused, the focus precision for the subject deteriorates. In view of this characteristic, the following adjustments may be made in order to perform the image processing considering the above correlations between the view and the subject.
- 1. When a distant view is focused, spatial filtering is performed so that a soft image is formed.
- 2. When a very close view is focused, spatial filtering is performed so that a sharp image is formed, and the level of UCR is intensified to improve the quality of the characters.
- In FIG. 4, the focal length information Dis read by the
film scanner 227 is input into theLUT 117. Color information COL to be developed in theprinter 200 is also input into theLUT 117, and various signals (UCR signal, FILTER signal, and SEN signal) are generated, as shown in FIG. 8, based on the color information COL and the focal length information Dis. - FIG. 8 illustrates the relationship between the input data and the output data in the
LUT 117. TheLUT 117 receives the focal length information Dis (two bits) read by thefilm scanner 227. The information Dis is defined as follows.Distant view: 3 Middle view: 2 Close view: 1 Very close view: 0 - The developing color COL indicating “1” represents a black color, while the developing color COL designating “0” represents a color other than black. Thus, by a combination of the color information COL and the focal length information Dis, the levels of the following types of signals are determined: the masking-coefficient control signal (UCR) to be input into the masking/
UCR circuit 109, the output filter coefficient control signal (FILTER) to be input into theoutput filter 111, and the 400-line/200-line switching signal (SEN) input into theprinter 200. The resulting signals are then output. - The UCR signal indicates the intensity level of the UCR operation: the intensity level decreases in the order of 0, 1, 2 and 3. The FILTER signal represents the enhancement level of an image edge: the edge enhancement decreases in the order of 0, 1, 2 and 3. The SEN signal “0” designates a 400-line signal, while the SEN signal “1” indicates a 200-line signal.
- FIG. 8 shows that the
LUT 117 sets the foregoing signals in the following manner when the image processing apparatus of this embodiment is under the standard condition. - 1. UCR processing is performed with an increasingly intensified level as a closer view is focused (in case where the developing color is black).
- 2. The sharpness of an image (FILTER signal) is increased as a closer view is focused.
- The number of recording lines is set at 200 indicated by the SEN signal “1” unless otherwise instructed by the user through the
operation section 1101. The masking/UCR processing circuit 109 generates a black signal K and performs masking according to the UCR control signal (UCR) output from theLUT 117. - The equations set out below illustrate masking/UCR equations performed in the masking/
UCR processing circuit 109. The minimum value MIN (C, M, Y) of C1, M1 and Y1 is first obtained to determine a black signal K1 according to the equation (2101). Then, 4×8 masking is performed by using the equation (2102) so as to output C2, M2, Y2 and K2. In this equation (2102), the coefficients m11 through m84 indicate masking coefficients determined by theprinter 200, while the coefficients k11 through k84 represent UCR coefficients determined by the UCR signal. - In the spatial filter processing section (output filter)111 shown in FIG. 4, two 5-pixel×5-pixel spatial filters are provided, and an output signal from one filter is connected to an input signal of the other filter. As the spatial filter coefficients, two types of smoothing (FILTER=3, 4) coefficients and two types of edge-enhancement (FILTER=0, 1) coefficients are provided and switched for each pixel in accordance with the level of the FILTER signal output from the
LUT 117. Further, by virtue of the two spatial filters, an edge enhancement operation is performed after a smoothing operation. It is thus possible to achieve edge enhancement with a decreased level of Moire fringing and also to output an image of higher quality by a combination of the two types of edge enhancement coefficients. - FIG. 9 is a flow chart illustrating the operation of generating a Dis signal indicating photographic information and a COL signal representing a developing color, the operation being performed by the
CPU 1102 of the imagesignal processing unit 209 according to an embodiment of the present invention. - In step S1, the user inputs through the
operation section 1101 the information representing a developing color (a color to be printed in the printer 200) and the number of lines for performing the PWM operation in theprinter 200. If the input developing color is monochrome, the COL signal is determined to be “1”. On the other hand, if the developing color is other than monochrome, the COL signal is determined to be “0”. The determined COL signal is then output to theLUT 117 in step S2. - In step S3, an image signal is input from the
film scanner 227. The image developed on thefilm 4 set in thefilm scanner 227 is read with theimage scanner 7, and the photographic data magnetically recorded on thefilm 4 is further read by themagnetic head 1. Subsequently, in step S4, the level of the Dis signal is determined as discussed above based on the input photographic data, for example, the focal length of the image when the image has been taken. Then, the determined Dis signal is output to theLUT 117 in step S5. According to this procedure, the UCR coefficient used in the masking/UCR circuit 109, the type of the spatial filter used in thespatial filter circuit 111, and the number of lines used for the PWM processing in theprinter 200 are determined. If it is found in step S1 that the 200-line mode (SEN 1) to be used in theprinter 200 has been instructed, theCPU 1102 does not particularly perform any operation. In contrast, if the 400-line mode (SEN=0) has been instructed, the SEN signal is directly changed to “1” and is output to theprinter 200. Alternatively, the information corresponding to “SEN=0” similar to the information shown in FIG. 8 is stored in theLUT 117, and the information is selected and directly output to theprinter 200. - In the foregoing embodiment, UCR processing and filtering processing are modified based on photographic data (the focal length in this embodiment) recorded on the film. However, the auxiliary data used in the present invention is not restricted to photographic information, and magnetic information or information in the form of a bar code recorded on part of a document image may be read with the
image scanner 201, and based on the read information, UCR processing or filtering processing may be performed on the document image. Moreover, the photographic data is not limited to the focal length, and various types of information concerning images may be used, or data which is directly usable for masking processing or UCR processing may be employed. - According to the foregoing embodiment, since the information is magnetically stored in a storage medium, the quality of a resulting image is not impaired. Also, the auxiliary data may be not only magnetically recorded, but may also be recorded by using a developer which may be only responsive to ultraviolet rays. The advantages offered by the present invention using a developer are similar to those exhibited by the present invention using the magnetically recorded information.
- The present invention may be applied to a system consisting of a plurality of devices (for example, a host computer, an interface, a reader, and a printer) or to a single device (for example, a copying machine or a facsimile machine).
- Further, the present invention may be achieved by the following modification. A storage medium on which a software program code for implementing the function of the foregoing embodiment is recorded may be installed in a system or a device, and the program code stored in the storage medium may be read and run by a computer (or its CPU or MPU) of the system or the device. In this case, the program code itself read from the storage medium achieves the function of the foregoing embodiment, and accordingly, the storage medium in which the program code is stored may form the present invention.
- The storage mediums for storing the program code may include floppy disks, hard disks, optical disks, magneto-optical disks, CD-ROMs, CD-Rs, magnetic tape, non-volatile memory cards, and ROMs.
- The invention also encompasses the cases where the function of the foregoing embodiment is achieved not only when the program code read from the storage medium is run by a computer, but also when an operating system (OS) run on the computer partially or wholly executes the corresponding processing in response to the instruction of the program code.
- The invention further encompasses the following case. The program code read from a storage medium may be written into a storage device provided in an extended-function board inserted into a computer or an extended-function unit connected to a computer. Then, the CPU provided in the extended-function board or the extended-function unit may partially or wholly perform the corresponding processing in response to the instruction of the program code, thereby implementing the function of the foregoing embodiment.
- As is seen from the foregoing description, the present invention offers the following advantages.
- An image signal is optically recorded on a recording medium, and at the same time, the corresponding auxiliary information is magnetically recorded on the recording medium. Then, a method for image processing performed on the image signal read from the recording medium is suitably determined based on the auxiliary information. Thus, optimal image processing is performed on the image input from the recording medium.
- Further, an image and the corresponding auxiliary information, such as focal length information, are recorded on a recording medium by different methods and are read therefrom. Based on the auxiliary information, the image is efficiently color-processed or filtered.
- While the present invention has been described with reference to what are presently considered to be the preferred embodiments, it is to be understood that the invention is not limited to the disclosed embodiments. To the contrary, the invention is intended to cover various modifications and equivalent arrangements included within the spirit and scope of the appended claims. The scope of the following claims is to be accorded the broadest interpretation so as to encompass all such modifications and equivalent structures and functions.
Claims (17)
1. An image processing apparatus comprising:
first reading means for reading an image recorded on a recording medium;
second reading means for reading focus information concerning a focus of the image when the image is taken, the image and the focus information being recorded on the recording medium by using different methods; and
color-processing means for color-processing the image read by said first reading means in accordance with the focus information read by said second reading means.
2. An image processing apparatus according to claim 1 , wherein the focus information comprises a focal length of the image when the image is taken.
3. An image processing apparatus according to claim 1 , wherein the image is optically recorded on the recording medium.
4. An image processing apparatus according to claim 1 , wherein the focus information is magnetically recorded on the recording medium.
5. An image processing apparatus according to claim 1 , wherein the recording medium comprises a film used in a camera.
6. An image processing apparatus according to claim 1 , wherein the color processing performed by said color-processing means comprises a masking or an under-color-removal processing.
7. An image processing apparatus according to claim 1 , wherein the focus information comprises at least an aperture, a shutter speed, and a type of the recording medium, all of which concern the image when the image is taken.
8. An image processing method comprising:
a first reading step of reading an image recorded on a recording medium;
a second reading step of reading focus information concerning a focus of the image when the image is taken, the image and the focus information being recorded on the recording medium by using different methods; and
a color-processing step of color-processing the image read in said first reading step in accordance with the focus information read in said second reading step.
9. A storage medium in which an image processing program is stored in such a manner that the program is readable by a computer, the program comprising:
a first reading step of reading an image recorded on a recording medium;
a second reading step of reading focus information concerning the focus of the image when the image is taken, the image and the focus information being recorded on the recording medium by using different methods; and
a color-processing step of color-processing the image read in said first reading step in accordance with the focus information read in said second reading step.
10. An image processing apparatus comprising:
first reading means for reading an image recorded on a recording medium;
second reading means for reading focus information concerning a focus of the image when the image is taken, the image and the focus information being recorded on the recording medium by using different methods; and
filter-processing means for filter-processing the image read by said first reading means in accordance with the focus information read by said second reading means.
11. An image processing apparatus according to claim 10 , wherein the focus information comprises a focal length of the image when the image is taken.
12. An image processing apparatus according to claim 10 , wherein the image is optically recorded on the recording medium.
13. An image processing apparatus according to claim 10 , wherein the focus information is magnetically recorded on the recording medium.
14. An image processing apparatus according to claim 10 , wherein the recording medium comprises a film used in a camera.
15. An image processing apparatus according to claim 10 , wherein the focus information comprises at least an aperture, a shutter speed, and a type of the recording medium, all of which concern the image when the image is taken.
16. An image processing method comprising:
a first reading step of reading an image recorded on a recording medium;
a second reading step of reading focus information concerning a focus of the image when the image is taken, the image and the focus information being recorded on the recording medium by using different methods; and
a filtering step of filtering the image read in said first reading step in accordance with the focus information read in said second reading step.
17. A storage medium in which an image processing program is stored in such a manner that the program is readable by a computer, the program comprising:
a first reading step of reading an image recorded on a recording medium;
a second reading step of reading focus information concerning a focus of the image when the image is taken, the image and the focus information being recorded on the recording medium by using different methods; and
a filtering step of filtering the image read in said first reading step in accordance with the focus information read in said second reading step.
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP9010309A JPH10210311A (en) | 1997-01-23 | 1997-01-23 | Image-processing method and device therefor |
JP010309/1997 | 1997-01-23 | ||
JP9-010309 | 1997-01-23 |
Publications (2)
Publication Number | Publication Date |
---|---|
US20020001407A1 true US20020001407A1 (en) | 2002-01-03 |
US6343144B2 US6343144B2 (en) | 2002-01-29 |
Family
ID=11746656
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US09/010,565 Expired - Lifetime US6343144B2 (en) | 1997-01-23 | 1998-01-22 | Method and apparatus for image processing, and storage medium |
Country Status (2)
Country | Link |
---|---|
US (1) | US6343144B2 (en) |
JP (1) | JPH10210311A (en) |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030205013A1 (en) * | 2002-05-03 | 2003-11-06 | Faus Group, Inc. | Flooring system having complementary sub-panels |
US20040009320A1 (en) * | 2002-05-03 | 2004-01-15 | Garcia Eugenio Cruz | Flooring system having complementary sub-panels |
US6720560B1 (en) * | 1999-12-30 | 2004-04-13 | Eastman Kodak Company | Method and apparatus for scanning images |
US20040074191A1 (en) * | 2002-05-03 | 2004-04-22 | Garcia Eugenio Cruz | Flooring system having microbevels |
US20040200165A1 (en) * | 2002-05-03 | 2004-10-14 | Faus Group, Inc | Flooring system having sub-panels |
US20060005498A1 (en) * | 2004-07-07 | 2006-01-12 | Vincente Sabater | Flooring system having sub-panels with complementary edge patterns |
US8875460B2 (en) | 1999-11-05 | 2014-11-04 | Faus Group, Inc. | Direct laminated floor |
US9172849B2 (en) | 2005-06-13 | 2015-10-27 | Thomson Licensing | Method and apparatus for color transformation by addressing a look-up table |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6533721B1 (en) * | 2001-07-27 | 2003-03-18 | Stryker Corporation | Endoscopic camera system with automatic non-mechanical focus |
US20040125211A1 (en) | 2002-09-03 | 2004-07-01 | Yoshirhiro Ishida | Image processing apparatus and image processing method |
Family Cites Families (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE3408107C2 (en) * | 1983-03-06 | 1996-05-30 | Canon Kk | Halftone color image recorder |
US5260732A (en) * | 1989-06-15 | 1993-11-09 | Nikon Corporation | Camera for producing a photographing field angle information and a printer for printing on the basis of photographing field angle information |
US5218392A (en) * | 1989-07-19 | 1993-06-08 | Fuji Photo Film Co., Ltd. | Apparatus for recording position information of principal image and photographic printing method using position information |
DE69024549T2 (en) * | 1989-07-25 | 1996-06-13 | Minolta Camera Kk | Image reproduction system |
JP2984306B2 (en) * | 1990-03-15 | 1999-11-29 | キヤノン株式会社 | Image reading device having focus adjusting device |
EP0532047B1 (en) * | 1991-09-12 | 1999-01-13 | Fuji Photo Film Co., Ltd. | Method of making photographic prints |
US5335082A (en) * | 1992-04-10 | 1994-08-02 | Opton Corporation | Method and apparatus for using monochrome images to form a color image |
JP3524108B2 (en) * | 1992-09-28 | 2004-05-10 | キヤノン株式会社 | Image processing device |
CA2144358C (en) * | 1994-03-15 | 2000-05-23 | Yuji Yamamoto | Film information communication apparatus, film information printing apparatus, information processing apparatus and index printer |
US5748287A (en) * | 1994-12-09 | 1998-05-05 | Konica Corporation | Photographic film reproducing apparatus using object brightness and exposure correction amount to develop photographed images |
-
1997
- 1997-01-23 JP JP9010309A patent/JPH10210311A/en not_active Withdrawn
-
1998
- 1998-01-22 US US09/010,565 patent/US6343144B2/en not_active Expired - Lifetime
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8875460B2 (en) | 1999-11-05 | 2014-11-04 | Faus Group, Inc. | Direct laminated floor |
US6720560B1 (en) * | 1999-12-30 | 2004-04-13 | Eastman Kodak Company | Method and apparatus for scanning images |
US20030205013A1 (en) * | 2002-05-03 | 2003-11-06 | Faus Group, Inc. | Flooring system having complementary sub-panels |
US20040009320A1 (en) * | 2002-05-03 | 2004-01-15 | Garcia Eugenio Cruz | Flooring system having complementary sub-panels |
US20040074191A1 (en) * | 2002-05-03 | 2004-04-22 | Garcia Eugenio Cruz | Flooring system having microbevels |
US20040200165A1 (en) * | 2002-05-03 | 2004-10-14 | Faus Group, Inc | Flooring system having sub-panels |
US7836648B2 (en) | 2002-05-03 | 2010-11-23 | Faus Group | Flooring system having complementary sub-panels |
US7836649B2 (en) | 2002-05-03 | 2010-11-23 | Faus Group, Inc. | Flooring system having microbevels |
US8112958B2 (en) | 2002-05-03 | 2012-02-14 | Faus Group | Flooring system having complementary sub-panels |
US8181407B2 (en) | 2002-05-03 | 2012-05-22 | Faus Group | Flooring system having sub-panels |
US20060005498A1 (en) * | 2004-07-07 | 2006-01-12 | Vincente Sabater | Flooring system having sub-panels with complementary edge patterns |
US9172849B2 (en) | 2005-06-13 | 2015-10-27 | Thomson Licensing | Method and apparatus for color transformation by addressing a look-up table |
Also Published As
Publication number | Publication date |
---|---|
JPH10210311A (en) | 1998-08-07 |
US6343144B2 (en) | 2002-01-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US5189523A (en) | Image processing apparatus | |
EP2096860B1 (en) | Image processing apparatus, image processing method, and computer program product | |
JP4280404B2 (en) | Image processing apparatus and method | |
US6343144B2 (en) | Method and apparatus for image processing, and storage medium | |
JP3176101B2 (en) | Image reading device | |
JP3333894B2 (en) | Image processing device | |
US7460268B2 (en) | Image processing device, image forming device, image processing method, image processing program, and recording medium containing the image processing program recorded thereon | |
US8928940B2 (en) | Image forming apparatus, image forming method, and integrated circuit | |
JPH11136532A (en) | Color image forming device and its control method | |
JP3407525B2 (en) | Image reproduction device | |
US5777750A (en) | Image processing method and apparatus which includes a color reversing function for a two-color image | |
JPH05316377A (en) | Picture processing unit | |
JP4497684B2 (en) | Image processing apparatus and method | |
JPH10155076A (en) | Color image forming device | |
JP2002223369A (en) | Image processing method, image processor and image forming device | |
US7375853B2 (en) | Image processing apparatus and method | |
JPH10322555A (en) | Image forming device | |
JPH0368270A (en) | Picture processor | |
JP3281391B2 (en) | Full-color image reproducing apparatus and full-color image reproducing method | |
JPH05344314A (en) | Picture processor | |
JP3079630B2 (en) | Image processing device | |
JPH10322490A (en) | Image forming device | |
JP2003134350A (en) | Image processing apparatus | |
JP2007019602A (en) | Image input output apparatus | |
JP3287612B2 (en) | Color image forming equipment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: CANON KABUSHIKI KAISHA, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:UCHIDA, YOSHIKI;REEL/FRAME:009161/0490 Effective date: 19980226 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
CC | Certificate of correction | ||
FPAY | Fee payment |
Year of fee payment: 4 |
|
FPAY | Fee payment |
Year of fee payment: 8 |
|
FPAY | Fee payment |
Year of fee payment: 12 |