GB2338160A - Simulating a brush stroke with varying coverage - Google Patents

Simulating a brush stroke with varying coverage Download PDF

Info

Publication number
GB2338160A
GB2338160A GB9922347A GB9922347A GB2338160A GB 2338160 A GB2338160 A GB 2338160A GB 9922347 A GB9922347 A GB 9922347A GB 9922347 A GB9922347 A GB 9922347A GB 2338160 A GB2338160 A GB 2338160A
Authority
GB
United Kingdom
Prior art keywords
data
ink
pixel
blurring
unit
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
GB9922347A
Other versions
GB9922347D0 (en
Inventor
Qinglian Guo
Kimiyuki Ohba
Yasufumi Sugawara
Tadahiko Mukai
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sony Corp
Original Assignee
Sony Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from JP7012109A external-priority patent/JPH08202336A/en
Priority claimed from JP7012110A external-priority patent/JPH08202893A/en
Application filed by Sony Corp filed Critical Sony Corp
Priority claimed from GB9601613A external-priority patent/GB2299249A/en
Publication of GB9922347D0 publication Critical patent/GB9922347D0/en
Publication of GB2338160A publication Critical patent/GB2338160A/en
Withdrawn legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • G06T11/20Drawing from basic elements, e.g. lines or circles
    • G06T11/203Drawing of straight lines or curves

Abstract

The present invention provides a method for forming a font of a character or a symbol including inputting a plurality of coordinate data lying on a contour of a stroke specifying a dash of the character or the symbol, preparing data of the stroke including a plurality of input coordinate data, having one of the input coordinate data as coordinate data specifying the beginning point of the stroke or as coordinate data specifying the end point of the stroke, and forming a font of the character or the symbol including the data of one or more of the strokes.

Description

2338160 METHOD AND APPARATUS FOR FORMING A FONT AND THE FONT PRODUCED
METHOD-AND APPARATUS FOR DRAWING A BLURRED.FIGURE This invention relates to methods and apparatus for forming a font and the font produced, and to methods and apparatus for drawing a blurred figure.
When displaying a character or a symbol on a display device using a computer, the character or the symbol is displayed on the display device based upon the font of the character or the s ymbo 1.
Such font may be exemplif ied by a bit map font def ined by a matrix of, for example, 24x24 dots, for displaying the character or the symbol by a set of these dots, or an outline font defined by plural points on a contour oil the character or the symbol and a curve interconnecting these points for displaying the character or the symbol by a curve interconnecting these points.
These'co-nventional fonts only contain data concerning the contour of the character or the symbol. The contour line of t he 1 Character or the symbol displayed based upon these fonts tends to be straightforward, that is, the portions inside and outside the contour line of the character or the symbol are unified in color, such as in black and white, respectively.
When the character or the symbol is actually written in the paper sheet with e.g., a writing brush. bleeding tends to be produced around the character or the symbol, or the character or the symbol tends to undergo blurring. Thus the character or the symbol is Such especially sheet with not necessarily straightforward in contour. bleeding or blurring represents crucial factors when a character or a symbol is written on a paper a writing brush. That is, when the character Qr the symbol is actually written on the paper sheet, the effect of blurring or bleeding is taken into account for more expressive representation of the character or the symbol.
If, when the character or the symbol is displayed on a display device using a computer, such blurred or bled character or symbol can be displayed, the character or the symbol can be displayed more expressively.
However, the conventional font consists only of data concerning the contour of the character or the symbol, as described above, such that the contour of the character or the symbol is drawn straightforwardly, with the inside and the outside ot the contour being unified in color, such as in black color and in white color. That is, there has up to now not been 2 known a f ont capable of displaying the blurred or bled character or symbol on the display device.
Since there lacks up to now a font capable of displaying -the blurred or bled character or symbol on the display device, there also are not known a method or apparatus for forming a font capable of displaying the blurred or bled character. or symbol on the display device.
In computer graphics for drawing a picture using a computer, threecolor data for color display and luminance data for monochromatic display are set as picture data for each pixel as a minimum picture unit, and a corresponding f igure, inclusive of characters and symbols, is displayed on a display device based upon the pixel-based picture data.
The simplest method in f igure drawing by such computer graphics is to f orm an area surrounded by a 1 ine and to draw a figure uniformly therein.
Since limitations are imposed on the f igure representation achievable with this method. a variety of f igure drawing methods have been proposed f o r enabling more expressive figure representation.
Among known f igure drawing methods, there is suc a method in which a figure is drawn based upon coordinated data entered by a pointing device, such as a mouse, as a coordinate data input device, 'so that the figure is drawn in synchronism with the movements of 'the pointing device for simulating -,.he figure 3 drawing by a writing brush. There is also a figure drawing method in which the figure is drawn in synchronism with the movement of the Painting device so that the color becomes thinner with progress in the figure drawing for simulating the writing with a writing brush and changes in color brought about by changes in the ink density and in the ink quantity contained in the writing brush. In addition, there is such a figure drawing method simulating the drawing with an air brush by decreasing the color density n a direction towards outside from a designated position and by increasing the color density as the drawing progresses.
Meanwhile, if a figure is actually drawn on the paper sheet using a writing brush containing ink, Indian ink (sumi) or water colors, collectively termed hereinafter as ink, and if the ink quantity contained in the writing brush is small, the writing brush carrying speed is high or the writing pressure is low, there are occasions wherein the ink is not sufficiently absorbed by the paper sheet and a portion of the f igure area is devoid of ink, thus producing blurring. Such blurring is an extremely important factor in expressiveness especially when writing a character or a symbol with a writing brush or when drawing a f igure with water colors, especial 1 y with Indian ink (sumi).
In a conventional method for representing such blurring on a displa device using a computer, it is assumed that the figure is drawn with a writing brush containing the ink, and the density 4 and the quantity of the ink contained in the writing brush are calculated so that the ink density and quantity are decreased as the figure drawing proceeds in order to produce blurring since the time when the ink density and quantity in the writing brush have become lower than pre-set values.
It is noted that. when a figure is actually drawn with a writing brush, blurring is produced at plural positions in the figure if the figure is of an increased length, the figure has several bends or when the writing brush carrying speed or the writing brush pressure undergoes variations in the course of figure drawing.
With the conventional blurred figure drawing method, since the density or the quantity of the ink contained in the writing brush is simp.ly decreased as the figure drawing progresses, so that only one blurred portion can be produced at an end portion of the figure and hence expressive figure representation cannot be achieved. In particular, a blurred character written as described above is significantly inferior in expressiveness to the character actually written with a writing brush.
There is therefore a need for a font capable of displaying the blurred or bled character or symbol on the display device, and a method and apparatus for forming such font.
In one aspect. the present invention provides a method for forming a font of a character or a symbol including inputting a plurality of coordinate data lying on a contour of a stroke specifying a dash of the character or the symbol, preparing data of the stroke including a plurality of input coordinate data, having one of the input coordinate data as coordinate data specifying the beginning point of the stroke or as coordinate data specifying the end point of the stroke, and forming a font of the character or the symbol including the data of one or more of the strokes.
In another aspect, the present invention provides an apparatus for forming a font of a character or a symbol including means for inputting a plurality of coordinate data lying on a contour of a stroke specifying a dash of the character or symbol, means for preparing data of the stroke including a plurality of input coordinate data, having one of the input coordinate data as coordinate data specifying the beginning point of the stroke or as coordinate data specifying the end point of the stroke, and means for forming a font of the character or the symbol including the data of one or more of the strokes.
In yet ano-,her aspect, the present invention provides a font of a character or a symbol wherein the character or the syribol is made up-of one or more strokes each representing a dash of a writing b rush and wherein the stroke data is made up of a plurality of coordinate data specifying 'the position 0,1 Z.he 6 stroke on the contour and coordinate data specifying the or the end paint of the stroke.
beginning point According to embodiments of the invention, a plurality of coordinate data lying on a contour of a stroke representing a dash of a character or a symbol are entered by coordinate input means. The stroke data forming means forms stroke data made up of plural input coordinate data entered by the coordinate input device, with one of the coordinate data being a coordinate data specifying a beginning point of the strake"or a coordinate data specifying an end point of the stroke. A character or a symbol i s f ormed f rom data of one of more st rokes f ormed by t he st roke data forming means.
Since the font contains the information concerning the writing sequence for the character or the symbol, it becomes possible to take into account the movement of the writing brush writing the character or the font and hence to display the blurred or bled character or symbol on a display device.
There is also a need for a method. and apparatus for drawing a blurred figure whereby two or more blurred portions can be produced at arbitrary positions in a figure in such a manner as to assure more representation.
In a further embodiment, expressive the invention provides a method for represent. i ng- a bl ur red f i cure on a d i spi ay devi ce based upon picture data set in a plurality of pixels in a figure. With the 7 npresent method, f i rst contour data including one or rno r e c oo r d i nat e po i nt S Ai ( i = 1, 2, 3,. - -) 1 y i n 9 o n a f i r s t c o n t o u r extending f rOM a beginning point to an end point of the figure and second contour data including one or more coordinate points B j (j = 1, 2, 3, -..) lying on a second contour extending from the beginning point to the end point are read. A pixel p lying on a line segment i k interconnecting a k'th coordinate Point Ak of the first contour data and a k'th coordinate point B k Of the second contour data is detected. It is then judged, based upon blurred area data specifying a blurred area of the f igure, whether or not the pixel P is within a blurred area. If the pixel P is within the blurred area, ink absence data specifying th-e ink depleted state is set as picture data in the pixel P. If the pixel P is not within the blurred area, ink presence data specifying the ink carrying state is set as picture data in the pixel P. A pixel Q on a line segment],,, interconnecting a (k+l)st coordinate point Ak+1 of the first contour data and a (k+l)st coordinate point Bk+1 Of the second contour data is detected. It is then judged. based upon the blurred area data, whether or not the pixel () is within the blurred area. If the pixel Q is within the blurred area, ink absence data specifying the ink depleted state is set as picture data in the pixel 0, whereas, if the pixel Q is not within the blurred area, ink presence data SPecifying the ink carrying state is set as picture data in the pixel Q. A pixel R lying on a line segment a n interconnecting the pixels P and 0 is detected. Picture data are set in the pixel R based upon respective picture data set in the pixels P and Q.
In another embodiment, the invention provides an apparatus for representing a blurred figure on a display device based upon picture data set in a plurality of pixels in a f igure. The apparatus includds means for reading f irst contour data i ncl udi ng one or more coordi nate poi nts Ai ( i = 1, 2, 3,...) lying on a first contour extending from a beginning point to an end point of the f igure and second contour data including one or more coordinate points Bj (j = 1, 2, 3,...) lying on a second contou r extend i ng f rom t he beg i nni ng poi nt to t he end po i nt and f irst pixel detection means f or detecting pixels on a line segment Ii interconnecting a coordinate point Aj of the first contour data and a coordinate paint B, of the second contour data based upon the first contour data and the second contour data suppliedfromthe data readout means. The apparatus also includes first picture data setting means for judging, based upon blurred area data specifying a blurred area of the f igure, whether or not the pixel detected by the f i rst pixel detection means is within a blurred area, and means for setting, if the pixel is within the blurred area, ink absence data specifying the ink depleted state as picture data in the pixel, and setting, if the pixel is not within the blurred area, ink presence data specif ying the ink carrying state as picture data in the pixel. The apparatus also 9.
includes second pixel detection means for detecting a Pixel positioned intermediate between the pixel lying on the line segment Ii set by the first pixel setting means and a pixel lying on the line segment 1 i+,. and second picture data setting means for setting Picture data on the pixel detected by the second pixel detection means based upon the picture data of the pixel lying on the line segment Ii detected by the first Picture data setting means and the Picture data of the pixel lying on the line segment lifl.
The invention will now be described by way of example with reference to the accompanying drawings, throughout which like parts are referred to by like references, and in which:
Fig-1 is a block diagram for illustrating a typical structure od a picture display system.
Fig.2 shows an example of a character.
Fig.3 illustrates a first stroke of the character shown in Fig.2.
Fig.4 illustrates a second stroke of the character shown in Fi 9. 2.
Fig.5 illustrates data for a stroke.
Fig.6---showsan example of data for the stroke shown in Fi 9. 3.
Fi 9. 4.
Fig.7 shows an example of data for the stroke shown in Fig.8 is a block diagram showing an illustrative font forming processor.
Fig.9 illustrates an example of input coordinate points and straight lines interconnecting these coordinate points.
Fig. 10 i I lustrates an example of input coordinate points and curved lines interconnecting these coordinate points.
Fig. 1 1 is a flow chart for illustrating the operation of the font forming Processor.
Fig-12 is a flow chart for inputting coordinate points as stroke data.
Fig.13 is a fiber structure Fig.14 is a diagrammatic view showing an example of fibers arrayed on a pixel 0 and neighboring pixels Pi.
Fig.15 is a block diagram showing an illustrative construction of a paper fiber structure data former.
Fig.16 is a graph showing an example of data representing the shape of fibers.
Fig.17 is a graph showing an example of data representing the position of fibers.
Fig-18 is a graph showing an example of data representing the direction of fibers.
Fig.19 is a block diagram showing an illustrative construction of a paper fiber distribution data &former.
Fig.20 is a flow chart for illustrating the operation of the paper fibe-r distribution data former.
Fig.21 is a diagrammatic view showing typical plural points arrayed at random by a dot arraying unit.
diagrammatic view showing an example of paper data 11 f-, Fig.22 is a diagrammatic view showing typical plural points uniformed in their positions by a uniforming unit.
Fig.23 is a diagrammatic view showing typical small areas formed by a small area former.
Fig.24 is a diagrammatic view showing typical f iber data arrayed at random by a random arraying unit.
Fig.25 is a diagrammatic view showing typical f ibers arrayed on a pixel.
Fig.26 is a block diagram showing an illustrative construction of a capillary tube calculating unit.
Fig-27 is a diagrammatic view showing an Mi number of fibers, with the angles of intersection between neighboring fibers being set at random.
Fig.28 is a flowchart showing the operation of a light transmittance calculating unit.
Fig.29 shows an illustrative figure presenting blurring.
Fig.30 illustrates an example of a figure processed by a blurring Processor.
Fig.31 illustrates an illustrative data structure of fullcontour data.
Fig.32 is a diagrammatic view showing an example of- a f igure drawing trajectory.
Fig.3-3 shows an example of a blurred area represented by the blurred area data.
Fig.34 shows an example of a blurred area represented by 12 blurred area data of Fig.32'.
is a diagrammatic view showing the data structure of Fig.35 the blurring parameters.
Fig-36 illustrates the relation between the blurring parameters and the blurred area.
Fig-37 illustrates interrupted blurring.
Fig.38 illustrates continuous blurring- Fig.39 is a diagrammatic view showing a data structure of shape data.
Fig.40 illustrates the curvature rj.
Fig.41 ia diagrammatic view showing an example of a data structure of picture data.
Fig.42 ia diagrammatic view showing an example of a data structure of writing brush data.
Fig.43 is a block diagram showing an illustrative construction of a blurring processor.
Fig.44 is a block diagram showing an illustrative construction of the writing brush data.
Fig.45 illustrates the state of ink quantity inputs for several points on a writing brush.
Fig.46 illustrates the state of ink density inputs for several points on a writing brush.
Fig.47 illustrates an example of the state of ink density inputs fo"r several points on a writing brush.
Fig.48 illustrates another example of the state of i n k 13 n.
density inputs for several points on a writing brush.
Fig.49 is a block diagram showing an illustrative construction of a data reader.
Fig-50 is a block diagram showing an illustrative construction of a blurring parameter setting unit.
Fig.51 illustrates an illustrative data structure of speed distribution data.
Fig.52 is a block diagram showing an illustrative construction of an ink data setting unit.- Fig.53 is a flowchart for illustrating the operation of a blurring parameter setting unit.
Figs. 54 to 57 are f lowcharts showing the continuation of the flowchart shown in Fig.53.
Fig.58 is a diagrammatic view showing an example of speed distribution data and writing brush pressure distribution data prior to conversion into bi-level data.
Fig.59 is a diagrammatic view showing an example of speed distribution data and writing brush pressure distribution data subsequent to conversion into bilevel data.
Fig.60 is a diagrammatic view showing an example of writing brush carrying data for producing blurring.
Fig.61 is a flowchart for illustrating the operation oil- a blurred area data former.
Fig'.62-is a flow chart for illustrating the operation of a contour data selector.
14 Fig.63 is a flow chart for illustrating the operation of an ink data setting unit.
Fig.64 illustrates an example of an area of ti = 1.0, 0.5 or 0.3.
Fig.65 is a flowchart for illustrating the operation of a first blurring parameter calculating unit.
Fig.66 is a flowchart showing the continuation of the flow chart of Fig. 65.
Fig.67 is a flowchart showing the continuation of the flow chart of Fig. 66.
Figs.68 to 87 illustrate examples of blurring for patterns 1 to 20, respectively.
Fig-88 shows an example of ink held in a gap between two fiber s in a pixel.
Fig.89 shows an example of an area occupied by ink held in a gap between two fibers in a pixel.
Fig.90 shows an example of ink held in a gap between a n number of fibers in a pixel.
Fig.91 is a schematic view showing a capillary tube.
Fig.92 is a block diagram showing an illustrative construction of a bleeding processor.
Fig.93 is a diagrammatic view showing an example of the number of pjxels and capillary tubes for illustrating the pixels connected by the capillary tube.
Fig.94 is a diagrammatic view showing another example of is I\ the number of Pixels and capillary tubes for illustrating the pixels connected by the capillary tube.
Fig.94 is a diagrammatic view showing an example of a branched point Fig.96 is of pixels from Fig.97 is set of Pixels pixel.
Fig.98 is a diagrammatic view showing still another example of a set of pixels, f rom a pixel Ai to an ink f low branching point pixel.
Fig.99 is a diagrammatic view showing an example of an area for producingbleeding.
Fig-100 is a diagrammatic view showing an illustrative construction of paper fiber structure data.
Fig.101 is a diagrammatic view showing an illustrative data construction of contour data.
Fig.102 is a diagrammatic view showing an illustrative data construction of current data.
Fig.103 is a diagrammatic view showing an illustrative data pixel.
diagrammatic view showing an example of a set pixel Ai to an ink f low branching point pixel.
diagrammatic view showing another example of a f rom a pixel A- to an ink f low branching point 1 construction of next data.
Fig..104 is a diagrammatic view showing an i 1 lustra.ive data construction of bled picture data.
Fig. 105 is a diagrammatic view showing an illustrative data 16 construction of luminance data.
Fig. 106 is a flow chart for illustrating the operation of a bleeding processor.
Fig. 107 is a flow chart showing the continuation of flowchart shown in Fig.106.
Fig.108 is a flowchart showing the continuation of the flowcharts shown in Figs.106 and 107.
Fig.109 is a f low chart showing the continuation of the flowcharts shown in Figs.106 to 109.
Fig-110 is a f low chart showing the continuation of the flowcharts shown in Figs.106 to 110.
Fig.111 is a diagrammatic view showing the method f o r setting bled Picture data in consideration of diffusion of ink particles on i-nk flow collision.
Fi 9. 1 12 i 1 1 ust rat es an exampl e of a cha racte r p resent i ng both blurring and bleeding.
Fig.113 illustrates the contour of a figure prior to antialiasing processing.
- Fig.114 illustrates the contour of a figure produced by aitialiasing by bleeding processing.
(1) Overall Construction of a Picture Display System Ref er.ring to the drawings, preferred embodiments of the method and apparatus for producing paper fiber structure data, paper f iber structure data, and a method and apparatus f or 17 drawing bled Pictures will be explained in detail. In the following embodiments, a picture display system is configured for displaying a picture made up of a set of pixels by displaying respective pixels based upon pixelbased picture data.
Referring to Fig. 1, the picture display system according to the present invention includes a data processor 10 for processing data by a central processing unit (CPU), and an input device 1 supplied with input data and conf igured for supplying the input data to the data processor 10. The picture display system also includes a Picture readout unit 2 for generating image data of a f igure and for supplying the image data to the data proc essor 10, and an internal storage device (memory) 3 for temporarily storing software or data required for data processing in the data processor 10 or storing data produced by the data processor 10. The picture display system also includes an external storage device 4 for recording software or data required for processing data in the data processor 10 and a display device 5 f o r displaying picture derived from the data processed by the da:a processor 10.
The data processor 10 includes a font forming processor 20 for forming fonts required for producing picture data of bled or blurred ch aracters or symbols based upon image data supplied from the input device 1 or f rom the picture readout device 2. and a paper f iber structure data forming processor 30 for forming paper 18 _\ 1 f iber structure data simulating the paper f iber structure. The data processor 10 also includes a blurring processing unit 40 for forming picture data of a blurred figure based upon data supplied f rom the input device 1, fonts formed by the font forming processor 20 and the paper fiber structure data formed by the paper fiber structure data former 30, and a bleeding processing unit 50 for forming picture data of abled figure based upon data supplied from the input device 1 and picture data formed in the blurring processing unit 40.
The input device 1 is comprised of a pointing device, such as a mouse, or a keyboard. The user enters data, such as coordinate data, parameters or commands, to the input device 1. The input data is thence supplied to the data proc essor 10. A plurality of such input devices may naturally be employed in combination.
Based upon these input data, the data processor 10 forms fonts required for forming picture data of bled or blurred characters or symbols in the font processing unit. 30. In addition, the data processor 10 forms picture data of a blurred figure or picture data of a bled figure in the blurring processing unit 40 or the bleeding processing unit so, respectively.
If ne-cessary, the data processor 10 reads out the software or data required for data processing from the external storage device 4 or writes data formed by data processing in the external 19 c storage device 4. In addition, the data processor 10 reads out the software or data required for data processing from the memory 3 or temporarily writes data formed by data processing in the memory 3.
That is, the memory 3 is conf igured for temporarily storing the software or data required for data processing by the data processor 10 or data formed by data processing by the data processor 10. If need be, the software or data are transmitted f rom the memory 3 to the data processor 10 or f rom the data processor 10 to the memory 3. The external storage device 4 is configured for storing the software or data required for data processing by the data processor 10, or data formed by data processing by the data processor 10. If need be, the software or data are r-ead from the external storage device 4 to the data processor 10 or f rom the data processor 10 to the external storage device 4. There is no particular 1 imitation to the recording medium employed as the external storage device 4 if it i s capab 1 e of st c r i ng t he sof twa re o r dat a reco rded t he reon. Thus t he reco rd i ng med i um may be enume rat ed by e. 9., a ha rd d i sc, a f lexible disc, a magneto-optical disc, an optical disc or a magnetic tape. A plurality of such recording media may naturally be employed in combination.
Examples of the software include a software for forming fonts required For forming picture data of bled or blurred characters or symbols, a sof tware for forming paper fiber structure data simulating the paper f iber structure, a software f or f orming picture data of a blurred f igure and a sof tware f or forming picture data of a bled figure.
Examples of the data required for data processing include ex i st i ng f onts or a var i et y of pa ramet e rs, wh i 1 e exampl es of dat a formed by data processing include a variety of data temporarily formed by the font forming processor 20, paper fiber structure data forming processor 30, blurring processing unit 40 or the bleeding processing unit 50, data of fonts formed by font forming by the font forming processor 20, paper fiber structure data formed by paper fiber structure data forming by the paper fiber structure data forming processor 30, picture data formed by blurring processing by the blurring processing u 1 nit 40- and picture data. formed by bleeding processing by the bleeding processing unit 50.
The data processed by the data processor 10 is displayed on the display device 5. The display device 5 displays characters or symbols derived f rom the fonts formed by the font forming processor 20, pictures derived f rom paper structure data formed by the paper f iber structure data forming processor 30, pictures derived from picture data formed by the blurring proces'sing unit 40 or pictures derived f rom picture data f ormed by the bleeding processing unit 50. There is no 1 imitation to the displaydevice 5 if it ' is- capable of displaying data processed by the data processor 10. Thus the display devi-ce 5 may be exemplified by 21 a display unit, printer or a plotter - A plurality of such display devices may naturally be employed in combination.
The p i ct u re readout clevi ce 2 i s conf i gu red f o r read i ng t he shape of a figure and transforming it into image data which is supplied to the data processor 10. The image data is employed for forming fonts in the font forming processor 20. That is, if it is desired to form fonts while reference is had to handwritten characters, one writes such characters on a paper sheet and image data of the characters is read into the picture readout unit 2 so as to be thence supplied to the data processor 10 for displaying a Picture derived f rom the image data.The user enters coordinate data, using the input device 1, such as a mouse. while having reference to the displayed picture. The font forming processor 20 then forms fonts using the coordinate data supplied thereto.
The data readout device 2 is conf igured for reading the shape of a f igure f rom outside and for supplying it to the data processor 10. Thus the data readout unit 2 may be omitted if there is no necessity of reading I'ie shape of the picture from outside. There is no particular limitation to the picture readout unit 2 if it is capable of transforming the sha-pe of the figure into image data. Thus an image scanner or a video camera, for exampl.e, may be employed as the picture readout device 2. Meanwhile, a plurality of such picture readout devices may be employed in combination.
22 Font Forming Processor Referring to Figs.2 to 12, a preferred embodiment of the font forming processor 20 is explained. In the following description, a stroke means a line of a character or a symbol. That is, a character MJ "+" shown in Fig. 2 is made up of a first stroke SR1 shown in Fig.3 and a second stroke SP2 shown in Fig.4. (2-1) Fonts
Referring to Figs.2 to 12, the font formed by the font forming processor 20 is explained.
When a character or a symbol is drawn on a paper sheet using a writing brush, bleeding around the character or symbol or the blurring in the character or symbol is produced depending on the manner in which the brush is carried and moved. Thus. for forming picture data realistically representing the character or the symbol as if they were actually drawn on the paper sheet with ink using the writing brush, it is necessary to scrutinize how the writing brush is moved in writing the character or the symbol.
The font formed by the font forming processor 20 based on the above considerations is a font having data concerning the stroke direction, and is made up of data of one or more strokes. The data oil each stroke is comprised of f irs+. contour data made up of p 1 u r a 1 coordinate points sequentially indicating the positions on a contour RN1 extending f rom a starting point PS up to an end" point PE of the stroke SR, and plural coordinate points sequential ly indicating the positions on a con- Lour RN2 extending 23 f rom the starting point PS up to the end point PE of the stroke SR, as shown for example in Fig.5. The first contour data is made up of coordinate points (x,, Y,) specifying the start point ps, n coordinate points (x,,, Y1,), (X12, Y12). - - (xln, Y1n) on the first contour RN1 and coordinate points (x., Y,) specifying the end point PE. The second contour data is made up of coordinate points (x,, Y,) specifying the start point PS,..a coordinate points (x21, y,,), (X22, Y22)... (X2n, y2n) on the f i rst contour RN2 and coordinate points (x e, Ye) specif ying the end point PE. It is noted that the values of n and m are integers not less than 0 and may be the same as or different from each other.
If, in the above described first contour data and the second data, coordinate points specifying the starting point PS is designated in-the preparation of stroke data, the so-designated coordinate points are set as the coordinate points (xsl ys) specifying the starting point. If no particular designation has been made, the coordinate points entered for the first time are set as the coordinate points (x,, Y,) specifying the star-Ling point. On the other hand, if coordinate points specifying the terminating Point PS are designated in the preparation of the stroke data, the so-designated coordinate points are set as the coordinate points (x., Y,) specifying the terminating point. If no particu-1ar designation has been made, the coordinate points entered for the last time are set as the coordinate points (x,, yd specifying the terminating point.
24 With such stroke data, since the coordinate points (x,, Y,) specifying the stroke start point and the coordinate points (x.. ye) specifying the stroke end point are set, the stroke direction is known. Thus it becomes possible to know the movement of a writing brush when writing a character or a symbol represented by a font made up of such stroke data. By preparing picture data of the character or symbol, based upon such font, a realistic expression may be achieved which will. give an impression as if the character or the symbol were drawn with the writing brush.
Meanwhile, if, when the character or the symbol is drawn with ink on the paper sheet using a writing brush, the sequence in which the character or the symbol is drawn, that is the stroke sequence, is -Changed, the sequence in which the strokes are overlapped is Changed, so that bleeding or blurring produced is also changed. Thus, if picture data is to be produced so that the character or the symbol is expressed more realistically for producing an impression as if the character or the symbol were drawn with t.he writing brush, it is necessary to consider the stroke sequence.
Thus, in the font formulated in the font forming -processor 20, the data of the respective strokes is arrayed in the writing sequence so that the stroke sequence will be known. It is noted that, if the data sequence is designated at the time of preparation of the stroke data, the stroke data sequence is the designated data sequence. If no'.narticular designation ha-S been Omade, the stroke data sequence is the ii-,,nut sequence. The stroke data sequence need not be set by arraying the stroke data in the writing sequence. For example, the stroke data may be set by formulating sequence data for specifying the sequence for the respective stroke data.
Taking an example of the letter MJ, which reads "+" shown in Fig.2, the above-described font is explained in detail.
The letter W which reads "+" is made up of the first stroke SR1 shown in Fig.3 and the second stroke SR2 shown-in Fig.4.
Referring to Fig.6, data of the first stroke SRI---is made up of. first contour data comprised of a coordinate point (xsl, YS1), specifying a start point PS1 of the first stroke SR1,'e.g., four points (xW Y11), (x12, Y,2), (xW Y13), (X14, Y14), and a coordinate point (xel, yel) specifying an end point PE1 of the first stroke SR1, arrayed in this order, and second contour data comprised of a coordinate point (xs2, Ys2), specifying a start point PS1, e.g., four points (x,l, Y,,), (x22, Y22), (X23, Y23), (X24. Y24), and a coordinate Point (xel, Yel). specifying the above terminal point PE1, arrayed in this order.
Similarly, data of the second stroke SR2 is made up of first contour data comprised of a coordinate point (X52, Ys2) specifying a start point PS2 of the first stroke SR2, e.g., four points (x3,, Y31), (x32, Y32), (x33, Y,3), (X34, Y,4), and a coordinate point (xel, Yel) specifying an end point PE1 of the second stroke 26 of data o (X51 ' YS1 making up SR2, arrayed in this order, and second contour data comprised of the coordinate point (x s2, y,2), spec i f y i ng t he above-ment i oned start point PS2, e.g., three points (X41, Y41), (X421 Y42) and (X4,, Y43), and the coordinate point (xe2, Ye2), specif ying the above terminal point, arrayed in this order.
Since the sequence of the strokes SR1 and SR2 of the letter mi which reads "+" is the sequence of the first stroke SR1 an the second stroke SR2, data of the f irst stroke SR1 and data of the second stroke SR2 are arrayed in the sequende of the data of the first stroke SR1 and the data of the second stroke- SR2.
Thus the font of the letter MJ which reads "+" is comprised f the first stroke SR1 made up of the coordinate points (xll, YM), (xl,, Y,2), (X13' Y13), (X141 Y,4) and (x.1, Y.1), the first contour data, coordinate points (x,l ' YSI) ' (x21, Y21), (x,2, Y,2), (x,3, Y23), (x24, Y24) and (xel, Yel), making up the second contour data, and data of the second stroke SR2 made up of the coordinate Points (xs2, Ys2), (x31, Y31), (X321 Y32), (x33, Y3,) and (xe2, Y,2) making UP the f i rst contour data and the coordinate points (x.2, y&, (X411 Y41)' (X42, Y32), (X43' Y43)' (X44' Y44) and (x,,, y& making up the second contour data. The data cf, the f irst stroke SR1 and the data of the second stroke SR2 are arrayed in this order.
Although both a set of the coordinate points specifying the starting point and another set of the coordinate points specifying the terminal point are set in the above-described font, at least one of these Sets Of the coordinate points suffices in order to comprehend the stroke direction. Thus it is possible to set only the coordinate point specifying the stroke start point or only the coordinate point specifying the stroke end point as the stroke data.
With the above-described font, the stroke data are constructed by the first contour data containing the coordinate point specifying the starting point and the coordinate point specifying the end point and the second contour data containing the coordinate point specifying the starting point and the coordinate point specifying the end point. However, -the stroke data may also be constructed by first contour data made up only of coordinate points on the first contour excluding the coordinate point specifying the starting point and the coordinate point specifying the end point and second contour data made up only of coordinate points on the second contour excluding the coordinate point specifying the starting point and the coordinate point specifying the end paint.
- Although the stroke data for the above font are made up only of the coordinate points, the stroke data may also have vector data. such as in the case of the conventional outline font, in addition to these coordinate points.
(2-2) Construction of Font Forming Processor An illustrative construction of the font forming processor 20 for preparing the above font is now explained.
28 Referring to Fig.8, the font forming processor 20 includes Oa stroke data forming unit 21 for forming stroke data in accordance with plural coordinate points entered by the input device 1, and a stroke sequence setting unit 22 for setting the sequence of data of plural strikes formed by the stroke data forming unit 21. The font forming processor 20 also includes an image data readout unit 23 for reading image data of the character or symbol from the picture readout unit 2 for writing in the memory 3 and an image data purveying unit 24 for reading out the image data from the memory 3 for supplying the image data thus read out to the display unit 5. The font former similarly includes a font readout unit 25 for reading out an existing font from the external storage device 4 for writing the read-out font in the memory 3 and a font purveying unit 26 for reading out the font from the memory 3 for supplying the readout font to the display device 5.
The st roke data f ormi ng unit 21 f orms, based upon plural coordinate points entered f rom the input device 1, the above mentioned stroke data, that is first contour data made up of the coordinate point (xsl YS) specifying the starting point of the stroke SP, n coordinate points (x,,, yl,), (X12, Y12), (xl., YM) on a f i rst contour RN1 and the coordinate point (xe, ye) specif yi ng the end poi nt of the st roke SR, and second contour data made up of the coordinate point (XS' YS) specifying the starting point of the stroke SP, n coordinate points (x21, Y2,), 29 (X221 Y22), - ' - (X2n, Y1n) on a second contour RN2 and the coordinate point (xe, Ye) specifying the end point of the stroke SR. Since each character or symbol is constituted by one or more strokes, the stroke data forming unit 21 forms data of one or more strokes for each font.
If Supplied from the input device 1 with data designating the coordinate point specifying the starting point, the stroke data forming unit 21 forms stroke data with the designated coordinate points as the coordinate point specifying the starting point. If not supplied f rom the input device 1 with data designating the coordinate point specifying the starting point, the stroke data forming unit 21 forms stroke data with the firstentered coordinate point as the coordinate point specifying the starting point. Similarly, if supplied from the input device 1 with data designating the coordinate point specifying the end point, the stroke data f orming unit 21 f orms stroke data with the designated coordinate point as the coordinate point specifying the end point. If not suppl ied f rom the input device 1 with data designating the coordinate points specifying the end point, the stroke data forming unit 21 forms stroke data with the lastentered coordinate point as the coordinate point specifying the end point.
For ease in entering the coordinate points on the stoke contour by the input device 1, the stroke data forming unit 21 includes display processing unit 211 for displaying the matter concerning the coordinate points and an editing processing unit 212 for editing the coordinate points.
For ease in understanding the relative position of the coordinate points entered by the inputting device 1, that is in understanding the shape represented by the stroke data, the display processing unit 211 displays points corresponding to coordinate points entered by the input device 1 on the display device 5. Also, for ease in understanding the sequence of the coordinate points entered by the input device 1, the display processing unit 211 displays a line interconnecting the coordinate points in accordance with the sequence of the coordinate points entered by the input device 1 on the display device 5.
Specifically, the display processing unit 211 displays points XY1, XY2, XY3, XY4, XY5, XY6, XY7, XY8, XY9, XY10, XY11, XY12, XY13, corresponding to coordinate points entered by the input device 1, on the display device 5, while displaying straight lines XY1-2, XY23, XY34, XY4-5, XY5-6, XY7-8, XY89 XY9-10, XY10-11, XY11-12, XY12-13, XY13-1, interconnecting these coordinate points, as shown for example in Fig.g. The display processing unit 211 displays the coordinate points entered by the input device 1 and the straight lines interconnecting these coordinate points on the display unit 5 in order to permit the user to comprehend easily the sequence of the input coordinate points.
31 Alternatively, the display processing unit 211 displays points XY1, XY2, XY3, XY4, XY5, XY6, XY7, XY8, XY9, -xylo, Xyll, XY12, XY13, corresponding to coordinate points entered by the input device 1, on the display device 5, while displaying smooth curves XY1-2, XY2-3, XY3-4, XY4-5, XY5-6, XY6- 7, XY7-8, XY8-9 XY9-10, XY10-11, XY11-12, XY12-13, XY13-1, interconnecting these coordinate points, on the display device 5, as shown for example in Fig. 10. Thus the display processing unit 211 displays smooth curves interconnecting the coordinate points entered by the input device 1 using a technique of interconnecting the respective coordinate points by curves such as splined cur"s or Bezier curves such as is used in the conventional outline fonts. This permits the user to know the shape of the stroke when the stroke data is complemented, as will be explained subsequently.
When displaying the po - ints corresponding to the entered coordinate points or the lines interconnecting the entered coordinate points on the display unit 5, the display processing unit 211 displays the coordinate points specifying the starting point of the stroke, the coordinate points specifying the end point of the stroke and the remaining coordinate points so as to permit facilitated visual discrimination between these different co6rdinate points. Specifically, the point corresponding to the coordinate point (xsl YS) specifying the stroke start point. the point corresponding to the coordinate 'point (x,, Y,) specifying the stroke end point and the remaining coordinate points (x,l, 32 t yn), ( x 2, Y12),. - - (Xln, Y 1n), (x,l, Y21), ( x2,, Y22) x 2.. 1 Y 1m) with different colors and/or luminance.
Although the display processing unit 211 displays dots and lines on the display device 5 only for data of a single stroke, it as also possible for the display processing unit 211 to display data of plural strokes simultaneously on the display device 5. That is, the display processing unit 211 displays only data f or a single stroke on the display device 5 or data of plural strokes simultaneously on the display device 5 in accordance with the instructions entered by the input device 1.
The editing processing unit 212 is responsive-to commands f rom the input device 1 to add coordinate data to the stroke data, move coordinate points of the stroke data one by one or collectively, or to delete the coordinate points from the stroke data one by one or collectively. If, by the operation of the display processing unit 211, dots and lines of data of plural strokes are displayed on the display unit 5, the editing processing unit 212 is responsive to commands form the input device 1 to select data of one of the strokes to effect processing on the data of the selected stroke. - The stroke sequence setting unit 22 then arrays the data of one or more strokes formed by the stroke data forming unit 21 in accordance with the stroke data sequence for setting the sequence cf the stroke data.
If supplied by the input device 1 with the stroke data 33 sequence designating data f rom the input device 1, the stroke 0 sequence setting unit 22 re-arrays the stroke data in the designated sequence for setting the stroke data sequence. If not s upplied by the input device 1 with the stroke data sequence designating data from the input device 1, the stroke sequence set fing unit 22 arrays the stroke data in the sequence in which the stroke data is formed in the stroke data forming unit 21 for setting the sequence of the stroke data.
It is unnecessary for the stroke sequence setting unit 22 to set the sequence of the respective strokes by arraying the data of the respective strokes according to the correc-t sequence. Fi r example, sequence data for specif ying the data of the respective strokes may naturally be formed for setting the sequence of the respective strokes.
By forming data of one o:r more strokes by the stroke data forming unit 21 and setting the data sequence of the respective strokes by the stroke sequence setting unit 22, the fonts may be formed which are comprised of data of one or more strokes and in which the data sequence of the respective strokes has been set.
If, when the coordinate points on the stroke contour are entered by the input device 1, the character or symbol corresponding to the desired font is displayed on the display unit 5, the coordinate points can be entered extremely easily. That is, the coordinate points can be entered extremely easily if the coordinate points are entered by designating the points 34 on the contour of the displayed character or symbol while the Ooperator has reference to the character or symbol displayed on the display unit 5.
The image data readout unit 23, image data purveying unit 24, image front readout unit 25 and the font purveying unit 26 are configured for inputting the coordinate points while the letter or symbol corresponding to the desired font is displayed on the display unit 5.
That is, the image data readout unit 23 reads out image data of the character or symbol corresponding to the desired font supplied from the picture readout device 2 for wri-ting in the memory 3. The image data purveying unit 24 reads out the image data f rom the memory 3 to transmit the read out data to the display unit 5. Since the character or symbol corresponding to the desired f ont is displayed in this manner on the display unit 5, the user is able to input the coordinate points while having reference to the character or font displayed on the display unit 5.
On the other hand, the font readout unit 25 reads out the font of the character or symbol corresponding to the desi red font supplied f rom the external storage device 4, and writes the readout -font in the memory 3. The font purveying unit 26 reads out the font f rom the memory 3 and purveys the read-out font to the display unit 5. Since the character or symbol corresponding to the desired font is displayed on the display unit 5, the user is abl e to i nput t he coo rd i nate poi nts whi 1 e havi ng ref e rence to the Ocharacter or f ont displayed on the display unit 5. The f ont pre recorded in the external storage device 4 may be any font provided an image of the character or symbol corresponding to the font can be displayed on the display unit 5. Examples of the fonts include conventional bit map fonts, outline fonts or fonts pre-formed by the font forming processor 20.
the illustrative operation of the font forming processor 20 is explained by referring to a flowchart shown in Fig.11.
At step ST-2-1-2, the font forming processor 20 refers to the character or symbol drawn on the paper for judging whether or not the font is to be formed. If the result is YES, that is if the font is to be formed, the font forming processor 20 transfers to step ST-2-1-2 and, if otherwise, the processor 20 transfers to step.ST2-1-3.
At step ST2-1-2, the font forming processor 20 displays the character or symbol drawn on the paper on the display unit 5 before transferring to step 2-1-5. Specifically, when the letter or symbol drawn on the paper is to be displayed on the display unit 5, the picture readout device 2 transforms the shape of the character or symbol drawn on the paper into image data and trarismits the resulting image data to the image data readout unit 23. The image data readout unit 23 writes the image data supplied form the picture readout unit 2 in the memory 3. The image data purveying unit 24 reads out the image data from the memory 3 and 36 transmits the readout image data to the display unit 5. The display unit 5 displays the picture derived from the image data purveyed from the image data purveying unit 24.
At step ST2-1-3, the font forming processor 20 refers to the font stored in the external storage device 4. for judging whether or not the font is to be formed. If the processor refers to the font (YES), the processor transfers to step ST2-1-4. If the processor does not refer to the font (NO), it transfers to step ST2-1 -5.
At step ST21-4, the font forming processor 20 displays the character or symbol corresponding to the font stored in the external storage device 4. The font forming processor 20 then transfers to step ST2-1-5. Specifically, if the character or symbol corresponding to the font stored in the external storage device 4 is to be displayed on the display device 5, the font readout unit 25 reads out the font from the external storage device 4 and writes the read-out font in the memory 3. The font purveying unit 26 then reads out the font f rom the memory 3 and sends the read-out font to the display device 5. The display device 5 displays the picture corresponding to the font purveyed from the font purveying unit 26.
- At step ST2-1 -5, t he st roke data f ormi ng uni t 2 1 of t he f ont forming processor 20 reads coordinate points entered by the user via the input device 1, as will be explained subsequently. The coordinate points read by the stroke data forming unit 21 are 37 plural coordi nate points (xsl YS). (X11, Y11), (X121 Y12), - - - (xl., Y1n), (x2P Y21), ( x22, Y22), - -. (X2n, Y1n), (xe 1 ye), as shown i n Fig.5. If the character or symbol drawn on the paper is displayed on the display unit 5, at step ST2-1-2, or if the character or symbol correspo- nding to the existing font is displayed on the display device 5 at step ST2-14, the user enters the stroke data via the input device 1, while having reference to the letter or symbol displayed on the display unit 5. The display processing unit 211 displays the points corresponding to the input coordinate points or lines interconnecting the input coordinate points on the display device 5 to permit facilit-ated visual recognition of the input coordinate points by the user.
At step ST2-1-6, the stroke data forming unit 21 forms, based upon the coordinate points read out at step ST21-5, the stroke data comprised of f irst' contour data made up of coordinate points (x,, Y,), (X11, Y11), (X12, Y12), - - - (Xln,yln), (xe. Ye), and second contour data made up of (xS1 ys). (X21, Y21), (X22, Y22),... (x2n,Yin), (xe, Ye). The leading coordinate points (x,, y,) of the first and second contour data specify the start point while the trailing coordinate points (x,, yd of the first and second contour data specify the end point. At this time, the stroke data forming unit 21 causes the dots and lines corresponding to the stroke data to be displayed on the display unit.
Next, at step ST2-17, the stroke data forming unit 21 judges, based upon the commands entered from the input device 1, 38 whether or not the formation of data of all of the strokes of a desired font has come to a close. If there is any data of the stroke yet to be formed (YES), the processor 20 reverts to step ST2-1-5. If the formation of data of all strokes has come to a close (NO), the processor transfers to step ST218.
If the Processor 20 reverts to step ST2-1-5 in order to produce data of a new stroke, the stroke data forming unit 21 causes the dots corresponding to the coordinate points of the data of the stroke formed previously and the lines interconnecting these dots to be displayed on the display unit 5. That is, if the user is desirous to check for-the balance between the array of the previously formed stroke and that of the stroke to be newly formed, the user has only to display the points corresponding to the coordinate points of the pre-formed stroke or the lines interconnecting these coordinate points on the display processor 211. If entry of the points corresponding to the coordinate points of the pre-formed stroke or the lines interconnecting these coordinate points on the display unit 5 leads to difficulties in entering the coordinaie points on the contour of a newly formed stroke, it is only sufficient if the points corresponding to the coordinate points of the pre-formed stroke or the lines interconnecting these coordinate points are not displayed as described above.
At step ST2-1-8, the stroke sequence setting unit 22 sets the sequence of data of one or more strokes formed at step ST2-1 39 6 by the stroke data forming unit 21. If supplied with the stroke data sequence designating data f rom the input device 1, the stroke sequence setting unit 22 re-arrays the stroke data in the designated sequence for setting the sequence of the data of the respective strokes. If not supplied with the stroke data sequence designating data f rom the input device 1, the stroke sequence setting unit 22 arrays data of the strokes in the sequence in which the stroke data has been formed in the stroke data forming unit 21 for setting the sequence of data of the respective strokes.
By the abovedescribed sequence of operations, the font forming processor 20 forms the font "+" which is formed by data of one or more strokes and in which the sequence of the data of the respective strokes is set.
When entering the coordinate points on the stroke contour at step ST2-1-5, the user enters the following coordinate points in accordance with the flowchart shown in Fig.12.
At step ST2-2-1, the user enters a command of initiating the entry of the coordinate points to the stroke data f orming unit 21, using the input device 1, and initiates th.e entry of the coordinate points on the strokecontour line. At step ST-2-2-2, Lhe" user enters the coordinate points (xsl Y,) specifying the stroke start point to the stroke data forming unit 21 using the input device 1. At step ST-2-23, the user sequential 1 y enters the coordinate points (xl,, Y11), (X12, Y12), - - - (xln, Yin) on the first contour line to the stroke data forming unit 21 using the Oinput device 1. Then, at step ST2-2-4, the user enters the coordinate points (xe, Ye) specifying the stroke end point to the stroke data forming unit 21 using the input device 1. At step ST-2-25, the user sequentially enters the coordinate points (xM, YM), (X22, Y22),... (x21, Y21) on the second contour line to the stroke data forming unit 21 using the input device 1. At step ST2-2-6, the user enters a command specifying the end of entry of the coordinate points to the stroke data forming unit 21 using the input device 1 to terminate the entry of the coordinate points on the stroke contour.
Thus, when entering the coordinate points at steps ST2-2-2 to ST2-2-5, the user displays the points corresponding to the input coordinate points and lines interconnecting the respective coordinate points on the display unit 5 using the display processor 211 as he or she enters the coordinate points on the stroke contour. If need be, the user adds, moves or deletes the coordinates, using the editing processor 212, for editing the coordinate points as he or she enters the coordinate points on the stroke contour.
(3) Paper Fiber Structure Data An illustrative example of the paper fiber structure data processing unit 30 is explained by referring to Figs.13 to 28. In the following description, a rectangular coordinate system is set in an area displaying a picture and pixels representing the
41 sma l 1 est Picture units are arrayed in each integer number coordinate (x, y) of the rectangular coordinate system. The region based on which the paper fiber structure data is to be formed is a region consisting of plural pixels on the rectangular coordinate system.
(3-1) Paper Fiber Structure Data The paper fiber structure data, formed by the paper fiber structure data forming processor 30, is first. explained.
Paper is constituted by plural interconnected fibers distributed in a reticular structure. The figure drawn on the paper with ink is formed by the amount of the ink stored between the fibers. The amount of the ink trapped between the fibers, that is the absorbed quantity of ink. depends upon the fiber structure. Thus, for forming picture data in which the figure drawn on the paper with ink is simulated more realistically, it is necessary to provided data simulating the paper fiber structure.
Since the ink flows through the fibers while the figure is drawn on the paper with ink, the ink flows through the fibers, so that bleeding is produced. The extent of the bleeding depends upon the gap between the fibers through which flows the ink. This gap is hereinafter referred to as a capillary tube. Thus, f o r formulating picture data in which picture data more realistically representing the figure drawn on the paper with ink, it is necessary to provide data simulating the capillary 42 tube of the Paper.
Based on this consideration, the paper f iber structure data, formed in the paper fiber structure data forming processor 30, is based on fiber data representing plural fibers constituting the paper which are arranged within an area made up of plural pixels. That is, the paper fiber structire data for the pixel P is comprised of a number M of f ibers passing through the pixel P, t he numbe rs M1 1 M21 M31 M41 M51 MG 1 M71 M8 of f i be r s pass i ng through the pixel P and also through eight pixels P], P2i P31 P41 P51 P61 P71 P8 neighboring to the pixel P, and the numbers of the capillary tubes S,, S2, S31 S41 S5 0 S61 S71 sa, as s hownf o r ex amp 1 e in Fig.13. The eight pixels P1, P21P31 PC P53P61 P71 P,, the eight numbers M,, M2, M39 M41 M51 M6) M71 MB and the eight numbers S1, S21 S31 S40 S5i S6i S71 sa are denoted as Pi, Mi and Si, respectively, where i.= 1,.2j 3, 4, 5, 6, 7, and 8. The numbers of f ibers M, the numbers of f ibers mi and the numbers of capillary tubes Si are represented by integers not less than 0. The numbe rs of f i be rs M, t he numbe rs of f i be rs Mi and t he numbe rs of capillary tubes Si are set for all pixels in the areas corresponding to the paper fiber structure data f-or constituting the paper fiber structure data.
The numbers of fibers Mi are the numbers of fibers passing through the pixel P and the pixels Pi. Thus, of the numbers Mi of the f ibers of the pixel P, those representing the f ibers passing through the pixels P,, P5j P and P, in this order are the 43 numbers M5 and M,, while the numbers of the f ibers M6 or M2 have no relevance.
With the paper f iber structure data simulating the f ine paper fiber structure, the absorbed ink quantity can be calculated since the numbers of fibers M are set. On the other hand, the numbers of capillary tubes Si formed in various directions from the pixel may be calculated since the numbers of fibers Mi are set, while the amount and the direction of the ink flowing from the pixel to the neighboring pixels can be calculated since the numbers of the capillary tubes Si are set. Thus, by employing such paper fiber structure data, -ink bleeding or absorption can be taken into account when forming picture data of a figure drawn on the paper with ink, so that it becomes possible to prepare picture data of a more realistically represented figure.
The capillar y tube is an ink conduit formed between the f i be rs, with the numbers of the capillary tubes Si being dependent upon the numbers Mi of the fibers. However, if the numbers of the capillary tubes are large, the width of each capillary tube is reduced, whereas, if the -numbers of the capillary tubes are small, the width of each capillary tube becomes larger. Thus, if the numbers of the capillary tubes are changed, the sum of the ink amount flowing in the capillary tubes is thought to be substantially unchanged. Thus, for simplifying the data, the number of the capillary tubes Si may be set to 1 44 and 0 for the numbers of f ibers Mi of not less than 2 and less than 2, respectively.
specifically, the paper fiber structure data is such that, if five fibers FIl, F12, FI3. F14 and F15 are arrayed for the pixels, P and Pi as shown for example in Fig.14, the respective valuesof the paper fiber structure datafor the pixel P are such that the numbers of fibers M, M1, M2P M31 M41 M51 M61 M7 and M, are 5, 1, 1, 2, 1, 2, 0, 3, respectively, and the numbers of capillary tubes S1, 521 S31 S4, S5, S61 S, and S, are 0, 1, 1, 0, 1, 0, 1, 0, respectively.
The paper f iber structure data may also comprise light transmittance I as data for representing the paper simulating the paper by the paper fiber structure data as a picture. The light t ransmi tt ance I i s set f o r r espect i ve pi xe 1 s and rep resent s 1 i ght transmittance for each pixel. The light transmittance I is a value which depends upon the number M of the fibers and becomes lower and higher with increase and decrease in the number of fibers M, respectively. By setting the luminance of respective pixels for the pixels encompassed by the paper fiber structure data based upon the light transmittance 1, and- displaying the luminance of the pixels on the display unit 5, the picture of the simulated paper may be represented by the paper fiber structure, thus permitting visual recognition of the state of the modelled or simulated paper structure data.
(3-2) Construction of the Paper Fiber Structure Data Forming Processor The illustrative structure of the paper f iber structure data forming processor 30 is now explained.
Referring to Fig. 15, the paper f iber structure data forming processor 30 includes a f iber data forming unit 31 for producing f iber data representing one of the f ibers making up the paper and a f iber distribution data f orming unit 32 for arranging the plural f iber data prepared by the f iber data f orming unit 31 in an area comprised of plural pixels f or forming f iber distribution data representing the fiber distribution. The -paper f iber structure data forming processor 30 also includes a ef iber number calculation unit 33 for calculating the numbers of f ibers M passing through the pixel P for each of the eight pixels Pi neighboring to the pixel P, and a neighboring fiber number calculating unit 34for calculating the numbers of fibers passing through the pixel P and also passing through the eight pixels Pi neighboringto the pixel P basedupon the fiber distribution data formed in the fiber distribution data forming unit 32. The paper fiber structure data formin: processor 30 also includes a capillary tube number calculating unit 35 for calculating the numbers of capillary tubes Si formed in the direction of the pixel Pi f rom the pixel P based upon the numbers of f ibers Mi calculated in the fiber number calculating nit 34.
The paper fiber structure data forming processor 30 forms paper fiber structure data, using the numbers of f ibers M 46 calculated by the fiber number calculating unit 33, the numbers Oof fibers Mi calculated by the neighboring fiber number calculating unit 34 and the numbers of capillary tubes Si calculated by the capillary tube number calculating unit 35 as paper fiber structure data.
The paper fiber structure data formed by the paper fiber structure forming processor 30 is stored by being written in the memory 3 or in the externa storage device 4. When a blurred figure is to be drawn by the blurring processor 40 or a bled figure is to be drawn by the bleeding processor, the paper fiber structure data is read out from the memory 3 or the external storage device 4 for forming picture data representing a figure drawn on the paper with ink.
Meanwhile, the fiber data formed by the fiber data forming unit 31 is the data representing one of the f ibers making up the paper, and is made up of data showing the f iber shape. f iber position data and fiber direction data.
The data specifying the f iber shape is a part of the cosine function as shown for example in Fig-6 and represented by the equation (3-1):
y = a x cos(x - b) where 0 S x s c, and a, b and are constants, p denoting the fiber length. It is noted that a represents the degree of bend of the fiber, such that, the smaller the value of a, the smaller is the bend of the f iber and, conversely, the larger the value 47 (D of A, the larger is the bend of thefiber. The values of p, _Q and _Q may be Pre-f ixed or may be varied within a pre-set range so as to be set at random depending upon fiber data. Alternatively, the values of A, _Q and _Q may be set to desired optional values entered by the input device 1 by the user or may be set at random depending upon the f iber data so to be as set to desired optional values entered by the input device 1 by the user.
By representing the f iber shape as a" part of the cosine function as indicated by the equation (3-1), it becomes -possible to represent the shape approximate to the actual paper fiber shape. However, the representation of the fiber shape is not limited thereto but may be represented by e.g., part of the sine function, or part of the n-ary curve, such as a quadratic or tertiary curve.
The data representing the fiber position is e.g., the coordinate value showing the end of a f iber (xl, Y1), while the data representing the f iber direction is the angle E) indicating the fiber arraying direction. The coordinate (xl, yl) and the angle e are values set at random by the f iber data forming unit 31.
The f iber data is obtained by transforming (x, y) of the equation (3-1) in accordance with the equations (3-2) and (3-3):
X X + xl...(3-2) y y + y 1... (3-3) 48 using the coordinate value (xl, yl) which is the fiber position representing data as shown for example in Fig.17, and also by rotating (x, y) the equation (3-1) n accordance with the equations (3-4) and (3-5):
X = X X cos(e) + y X cos(e)...(3-4) y = -x x sin(e) + Y x cos(e)... (3-5) using the angle e which is the data representing the fiber direction, as shown for example in Fig.18.
These plural fiber data, formed by the fiber data forming unit 31. are arranged in an area for forming the- paper fiber structure data by the fiber distribution data forming-unit 32 for forming fiber distribution data.
For taking int account both the uniform distribution and partial distribution of the paper fibers, the fiber distribution data forming unit 32 includes a dot arraying unit 321 f o r arraying plural points at random in an area made up of plural pixels encompassed by the paper fiber structure data and a uniforming unit 322 for uniforming the plural points arranged by the dot arraying unit 321 by relaxation processing, as shown in Fig.19. The f i be r distribution data forming unit 32 also includes a sub-area forming unit 323 for forming plural sub-areas centered about each dot uniformed by the uniforming unit 322 and a random arraying unit 324 for random arraying the same number of fiber data in each of the subareas formed by the sub-area forming unit 323.
49 For producing the f iber distribution data in the f iber Odistribution data forming unit 32, plural points are arranged at random in an area made up of plural pixels encompassed by the paper fiber structure data by the dot arraying unit 321. The plural points arranged at the dot arraying unit 321 are uniformed by relaxation processing by the uniforming unit 322. Then, a plurality of sub-areas, each centered around each point uniformed by the uniforming unit 322, are formed by the sub-area forming unit 323. The same number of f iber data ar e arranged at random in each sub-area formed by the sub-area forming unit 323 for forming fiber distribution data.
Referring to the flowchart of Fig.20, an illustrative operation of the fiber distribution data forming unit 32 is explained.
At step ST3-1, the dot arraying unit 321 arrays at random n dots TN in an area RI made up of plural pixels encompassed by the paper fiber structure data, as shown for example in Fig.21. The value of n, which is an integer not less than 2, may be preset or entered at a desi red value by the user via the input unit 1.
Then, at step ST3-2, the dot arraying unit 321 calculates the radius R, of the sub-area formed within the area RI made up of plural pixels encompassed by the paper fiber structure data. The value of the radium R, may be pre-set, entered at a desired value by the user via the input unit 1, or calculated based upon the value of the number n of the dots arrayed at step ST3-1 - if the value of the radius R, is to be calculated based upon the number _n of the dots arrayed at step ST3-1, it is calculated -in accordance with the equation (3-6):
R20 = 2.0 x (X x Y/n)112... (3-6) where X and R denote the transverse length and the vertical length of the area RI made up of plural pixels encompassed by the paper fiber structure data. Then, at step ST-3, the uniforming unit 322 carries out relaxation processing for moving the respective dots for uniforming the positions of the n dots arranged at the step ST3-1.
As an example of the relaxation processing, the relaxation processing on a point TNa having the current position (X, Y) is explained.
At step ST3-3-1, the uniforming unit 322 detects m dots TNj (j = 1, 2, 3,. ..m), included within a circle having a radius RO centered about (X, Y), from among the dots TN placed at random at step ST3-1. The radi us R, i s t he val ue f ound at step ST3-2. It is noted that m<n and a point TNa is not included among the dots TN j- At step ST3-3-2, the uniforming unit 322 finds _i vectors (Xj - X, Yj - Y) proceeding f rom a point TNa to a point TNj.
At step ST3-3-3, the uniforming unit 322 finds the sum (Xfl Yf) Of v>ctors (Xj X, Yj - Y) proceeding f rom the point TN, to the point TNj, where (X, Yt) is represented by the following t 51 equation (3-7):
(Xf Yf) = (X, - X. Y1 - Y) + (X2 X, Y2 - Y) + ( X11) X, Y 0 - Y) Then, at the position the equation X' X + d X X f Y' Y + d X Y f .. (3-7) step ST33-4, the uniforming unit 322 transforms of the pointTNa (X, Y) into (X', Y') asshown by 3-8) and (3-9):
..(3-8) ..(39) where _d is a pre-set proportionality constant.
The relaxation processing on the dot TN, is performed by the above step ST3-3-1 to ST3-3-4. Thus, by sequentially performing the steps ST3-3-1 to ST33-4 for 11 dots TN, the n points TN is relaxed, so that the n points TN placed in the area RI made up of plural pixels encompassed by the paper fiber structure data are uniformed, as shown for ixample in Fig.22.
The relaxation processing is not limited to the above-given processing provided the positions of plural points may be uniformed. Thus, any of known relaxation processing operations may be employed.
If the step ST3-3 comes to a close, the uniforming unit 322 judges, at step ST3-4, whether or not the relaxation p-rocessing is to be iterated. The number of the relaxation processing operations_needs only to be changed depending upon the state of distribut,on of the fibers constituting the desired paper. if the distribution of 't he f ibers making up the paper exhibits 52 partiality over the entire area, the number of iteration of the relaxation processing is diminished, whereas, if the paper exhibits partiality to a lesser extent over the entire area, the number of iteration is decreased. The number of times of iteration of the relaxation processing may be pre-set or entered via the input device 1 so that the number of times of the iteration may be freely set by the user. If the relaxation processing is to be repeated (YES), the processing reverts to step ST3-3, whereas, if the relaxation processing comes to a close (NO). the processing returns to step ST3-5.
At step ST3-5, the sub-area forming unit 323 forms n circular sub-areas MR, each centered about one of _n uniformed dots TN, as shown for example in Fig.23. The value of the radius RO of the sub-area MR is one found at step ST3-2.
Then, at. step ST3-6, the random arraying unit 324 places at random f iber data in the respective subareas MR so that the same number of f ibers F1 as that of the smal 1 sub-areas MR wil 1 be arranged in the small sub-areas MR as shown for example in Fig.24.
BY the above steps ST3-1 to ST3-6, the f iber data formed by the fiber data forming unit 31 are placed with uniform distribution and partial distribution in the area for forming paper fiber-structure data- If tle f iber data are formed in this manner, it becomes possible to realize the -Fiber distribution data having both the 53 0 uniform f iber data distribution over the entire area and the partial fiber data distribution over a small-sized area. That is, the f ibe-r data may be distributed so as to have part. ial distribution in a smallsized area by uniforming the positions of the dots by the uniforming unit 322 so that partial dot positions are left to a small extent. On the other hand, the fiber data may be distributed so as to be uniform over the entire area by setting the same number of fibers as the respective subareas formed by the small area forming unit 323 in these sub areas.
Based upon the thus formed fiber distribution data, the fiber number calculating unit 33 finds the number of f-i bers associated with the pixel P. That is, the fiber number calculating un-it 33 finds the number of fibers M passing through the pixel P based upon the f i - ber distribution data formed by the f iber distribution data forming unit 32. This number of the f ibers M is found f or each pixel so as to be the paper f iber structure data. Specif ical ly, if f ibers FI1, FIZ, FI3 and F!, are arranged for the pixel P, as shown in Fig. 25, the number of fibers M of the pixel P is equal to 4. on the other hand, the neighboring fiber number calculating unit 34 finds the number of the fibers for the pixel P according to the fiber directions.
Thus the unit 34 finds the numbers of fibers Mi passing through the pixel--P and through the eight pixels Pi neighboring!c the pixel P based upon the fiber distribution data formed by the 54 fiber distribution data forming unit 32.
If the fiber passes through the pixel P and through two or more pixels of the eight neighboring pixels P, neighboring to the pixel P. it is then assumed that the f iber is arrayed in the direction of the f iber P, which the f iber emanating f rom the pixel P f i rst traverses and thus the number Mi of the f ibers for the pixel Pi traversed first by the fiber emanating from the pixel P is incremented by one. Specifically, if the fiber traverses the pixels P,, PS, P, P11 P, in this order, the numbers of fibers M5 and M1 are incremented by one, while the numbers of fibers M6 and M2 are not incremented by one.
The numbers of fibers Mi are found from pixel to pixel to pixel in order to provide the paper fiber structure data. specifically, _if the five fibers F1P F12, FI3, FL, and FI5 are arrayed for the pixels P and' Pi, as shown in Fig.14, the number m of fibers of tIhe pixel P is 5, while the numbers of the fibers MJ, M21 M31 M41 MS, M61 M7 and M8 become equal to 1, 1. 2, 1, 2. 0, 3, 0, respectively.
The capillary tube number setting unit 35 finds the numbers of the capillary tubes for the pixel P according to directions. That is, the calculating unit 35 finds the numbers of capillary tubes Si formed in the respective directions of the eight pixels neighboring to the pixel P based upon the numbers Mi of fibers as calculate'd-by the neighboring fiber number calculating unit 34. The numbers Si of the capillary tubes are found on the pixel 0 basis to Provide the paper fiber structure data.
Since the ink duct formed in the f iber gap is the capillary tube, the number of the capillary tubes Si is dependent upon the number of fibers Mi. However, if the number of capillary tubes is large, the capillary tube is reduced in width, whereas, if there is only a small number of the capillary tubes, the capillary tube is increased in width. Thus it may be presumed that the sum total of the ink quantity flowing through he capillary tubes remains substantially constant even if the number of the capillary tubes is changed. Thus the number of capillary tubes Si may be reduced to data specifying only the presence or absence of the capillary tubes for data simplification. Specifically, the number of the capillary tubes Si may be set to 1 or 0 according as the capillary tubes are or are not constituted by Mi fibers, respectively.
In deciding w hether or not the capillary tubes are constituted by Mi tubes, it is assumed that the capillary tubes are formed for the number of fibers Mi equal to 2 or more while the capillary tubes are not formed if the number of fibers M i S 1 less than 2. That is, the number of capillary tubes S, is set to 1 or 0 for the number of f ibers Mi equal to not less than 2 or less than 2, respectively. Specif ically, if the f iber f ibers FI,, FIZ, F13, FL4 and F15 are arrayed for the pixel P and the pixel Pi as shown-i-n Fig.14, the numbers of the capillary tubes S] S21 S31 S41 SS, S6, S7, SaareO, 0, 1, 0, 1, 0, 1, 0, respectively.
56 For more accurate simulation of the paper structure, the angles of intersection of the M1 f ibers may be taken into account in setting. the numbers Si of the capillary tubes. The reason is that, if the angle of intersection of the nei ghboring f ibers is large, the f iber gap is increased so that the f iber gap ceases to function as the capillary tube and hence the ink cannot f low through the f iber gap.
If the number of the capillary tubes Si is to be set in consideration of the angles of intersection of the M 1 f ibers, the capillary tube calculating unit 35 is made up of an intersection angle setting unit 351 for setting the angles of intersection between neighboring f ibers at random, and a capillary tube number setting unit 352 for setting the numbers of capillar'y tubes Si, as shown for example in Fig. 26. After setting the angles of intersection between neighboring f ibers of Mi f ibers in the direction of the pixels Pi neighboring to the pixel P at random by the intersection angle setting unit 351, the capillary tube number setting unit 352 sets the number of the capillary tubes Si to 1 or 0 if at least one of the intersection angles as set by the intersection angle setting nit 351 is within a preset angle or if the totality of the intersection angles as set by the intersection angle set-Ling unit 351 is not within the pre-set angle, respectively. Specifically, if the number M1 of fibers of the pix'e-1 P is 5, the 'intersection angle setting unit 351 sets the intersection angles Ae,, ie,, AE), and Ae, between neighboring 57 99 goucliiwsueii L4 6 1 GLI 1 O UOLlleiedo;a^ i zle j I sn 1 eq_L L G X i d aqi;46no-jq,-, aouelliwsue-j), lqELL aqi 5GWO0aq jasiel aq), 1V4 s-iaq; awi;o jeqwnu aq), jassel a4I PulE LGxid 9t43. wEno-iLID, eouielllwsueil lw5L1 a q-1 sawooz?q ssal GqI v4 sjeqp aqi 4.o jeqwnu aqi ajow GL;I leql os laxid qoee O v4 sjaqp O jaqwnu eWl uodn pasRq palelnoLt>D SL 1 GOUCIZILtSUle.)), 1q611 aq_L 'leirp a.,n5.onj),s iaq jaded @t43, se 1 @OuCIZLwsuEJ), 1WELL PalelnolE!o aqi sesn pule V4 sjaq o jeqwnu aLli uOdn pas-eq 1 aouellwsuie-jl Zt4611 paseq-1exLd awl sazelnolú>o ge lun Euilelnoleo aoucilwsuciZ lqBil aW_1 CC Zpun 6uiZelnoLE0 jaqwnu jeqp; aqi Aq palelnoleo sc IN sjaqp. O Jaqwnu awl uodn Pasleq p@:Dield sL jeqp awl Woiwm ui lgxid at41 ui a:DuLZltLsuejl 1q5L1 Gwl BUI1f201PUL 1 Goulellwsuejl lq6iL pasieq-laxid aqZ 6uipui -jo qú lun buizelnoleo eouielliwsuejl 1q611 le sapnJOUL OC jossaooid EuLwjo eziep a-inZon-jls jeqp.1 j;Ddled at4z 'elep ein),oniis Jaq jadled awl SE? leS OSIC SL 1 G0UleZZLWSUleil 1t461L 1 alEue IGs-aid awl ULqZM IOU SL[SC lun 6ULlIGS 916UC UOLIOaS.JalUL 9LI:L Aq ZGS SE 5916UE UOL1095JeJUL GqI O AIL1C101 awl JO GLEuE lasaid c upwim si L95 1Lun 6ULIZaS GLEuE U01l0GE.JG1UL GqZ Aq Z@S SE tGV Pule teV 'ZE)V '1E)V Sal6Ule uoiloE)sjelui aw), o euo isfal iE; 0 -io L ol.!S saqnl. AjielldeD G1,4Z 4o _jeqwnu et4z slas zGC liun butz-.1as -jGqwnu aqnl AielldCO G141 LZ-61-1 ul UmOWS sC ILSC lun 6uillas G16UE UOLIOG5-JGZUL GLJ", 12 Wopulei lc sij Pue t-i 'Cii 'Zii 'I ii sjaq GAL: owl;o euo calculating unit 36 will be explained by referring to the flowchart of Fig.28. In the following explanation, it is assumed that the larger the value of the light transmittance I, t_he lesser becomes the light transmittance and the smaller the value of the light transmittance I, the higher becomes the light transmittance.
* At step ST3-11, the light transmittance calculating unit 36 sets the value of light transmittance I to (d x e) as an initial value. where _d and e are pre-set constants, with 0 < d < 1..
At step S73-12, the light transmittance calculating unit 36 sets the value of the f lag for the number of times of processing operations f to 0.
At step ST3-13, the light transmittance calculating unit 36 i nc rements t he. val ue of t he f 1 ag spec i f yi ng t he numbe r of t i mes of processing operations to 1 for setting the number of times of the processing operations.
At step ST3-1 4, t he 1 i 9 ht t ransmi -ttance cal cu 1 at ing un i t 36 compares the values of f and M to each other for repeating the processing a number of times equal to the number M of the f ibers for the pi.xel P. The light transmittance calculating unit 36 transfers to step ST315 or terminates the processing-if f 5 M or otherwise. respectively.
At step ST3-15, the light transmittance calculating unit 36 sets the",,alue of the light transmittance I to (d x e + (1 d) x 1) as a change in the light transmittance caused by a sole 59 f iber. The 1 ight transmittance calculating unit 36 then reverts to step ST313 for iterating the processing a number of times equal to the-number of fibers m for the pixel P.
By setting the luminance of the respective pixels based upon the thus set light transmittance I and displaying the luminance in the display unit 5, the picture of the paper simulated by the paper fiber structure data may be displayed such that the state of the simulated paper fiber structure data may be checked by visual inspection by the user. (4) Blurring Processor Referring to Figs.29 to 87, an illustrative
embodiment of the blurring processor 40 is explained n detail.
If. while the writing brush is moved from a sta'r-Ling point towards an end. point for drawing a picture, the amount of ink contained in the brush is small, the brush carrying speed is high or the brush pressure is low, a region depleted of ink, referred to hereinafter as a blurred area KS, is produced as a result of blurring. The blurring processor 40 mainly forms picture data for displaying a blurred figure. Such picture data is supplied to the display unit 5 for drawing a blurred figure on the display surface of the display unit 5.
The figure processed by the blurring processor 40 is a figure ZK surrounded by a first contour rnl interconnecting plural pdi-nts pa inclusive of a starting point ps and an end paint pe and a second contour rn2 interconnecting plural points pb inclusive of the starting point ps and the end point Pe. as shown for example n Fig.30. Specifically, the figure processed by the blurring processor 40 is a figure represented by stro.ke data of a font formed by the above-mentioned font forming processor 20.
The basic processing performed by the blurring processor 40 is to set ink present data specifying the ink present state as picture data for pixels of the figure within a non-blurred region, referred to hereinafter as a drawing region, and to set ink absent data specifying the ink absent state as picture data for pixels within the blurred area KS for forming picture data for representing a blurred figure ZK.
(4-1) Data Employed in Blurring Processing First, full-contour data, blurred region data, blurring parameters, shape data, picture data and writing brush data, which are main data employed in the blurring processing by the blurring processor 40, are explained.
(4-1-1) Full Contour Data The full-contour data is the data specifying the contour of a figure and the drawing direction. For example, the fullcontour data is comprised of first contour data including a contour interconnecting a start point and an end point of a figure and the drawing direction, and second contour data includind -another contour interconnecting the start point and the end point oil the figure and the drawing direction, as shown in 61 0 Fig. 31.
The f irst contour data is an array of m coordinate points Ai (i = 1, 2,_ 3,..., m), M being an integer not less than 1, -on the f irst contour of the f igure, arrayed in the f igure drawing direction. as shown in Fig.31. The f irst coordinate paint A, and the last coordinate point AM represent the starting point and the end point of the f igure, respectively. Similarly, the second contour data is an array of n coordinate points Bj Cj = 1. 2, 3, n), a being an integer not less than 1, on the second contour of the figure. arrayed in the figure drawing direction, as shown in Fig.38. The f irst coordinate point B1 and the last coordinate point B. represent the starting point and the end point of the figure, respectively. Thus the coordinate points A, and B1 represent the same point, while the coordinate points A. and B,, represent the same point.
The above-described fullcontour data is complemented as explained subsequently so that coordinate points are added to the f irst contour data and the second contour data for further smoothing the contour of the figure. Thus it suffices to set the initially set first and second contour data so as to permit the approximate contour of the figure to be comprehended. Th e coordinate points are added to the first and second contour data so that the numbers of the coordinate points of the f i rst. contour data m arTd- the numbers of the coordinate points of the second contour data n will be equal to each other by way of 62 complementation, as will be explained subsequently. Thus it is unnecessary for the initially set f irst and second contour data to be equal to each other.
Specifically, these full-contour data are stroke data of the font formed by the font forming processor 20. Thus the blurring processor 40 performs blurring processing on the stroke data of the font formed by the font forming processor 20.
If there exist plural f igures to be drawn, for example, if a character or a symbol is comprised of plural strokes each of which is a f igure to be drawn, plural sets of full-contour data are provided for coping with the respective strokes. The respective sets of the full-contour data are set so that the full-contour data will be arrayed in the drawing sequence of the respective fi-gures to permit the drawing sequence of the respective figures to be understood. Specifically, the plural sets of the full-contour data is a font comprised of data of plural strokes formed by the font forming processor 20. That is, in the case of the font comprised of data of plural strokes formed by the font forming processor 20, the d a -1 a of the respective strokes each corresponding to the f ul 1 contour data are arrayed in the writing sequence for setting the data sequence of the respective strokes. The blurring processor 40 performs blurring processing on the font formed by the font forming processor-20.
(4-1-2) Blurred Region Data 63 The blurred region data is the data representing the position of the blurred region and is comprised of data representing- the length of m trajectories along the f i rst -or second contours extending within the f igure f rom the starting point to the end point. These trajectories are hereinafter referred to as drawing trajectories. The position within the f igure of the blurred area is represented by the lengths of these m trajectories.
The drawing trajectories E3K are M lines extending from the start point ps as far as the end point pe and spaced apart with an equal spacing from each other along the width of the figure ZK as shown for example in Fig. 32. Both end drawing traject ories are equivalent to the first contour line rnl and the second contour line rn2 of the figure ZK. Although the number M of the drawing trajectories is small in Fig.32 for ease in understanding, the number m of the drawing trajectories EK is actually much larger for representing the blurring more precisely.
The blurred area data is comprised of m blurring start length data on the m drawing trajectories BK beginning from the start point ps and terminating at blurring start points KPS and m blurring end length data on the m drawing trajectories E3K beginning f rom the start point ps and terminating at blurring end points OE---; thereby specifying the position in the drawing ZK of the blurred region KS.
64 If there exists no blurred area KS on a given drawing trajectory BK, the blurring start length data and the blurring end length data on the drawing trajectory are set to an equal value. Specifically, the blurring start length data and the blurring end length data on the drawing trajectory are both set to 0 or 1.
The blurring start length data is a value comprised within a range of 0 and 1, obtained on normalizing the length from the start point ps as f ar as the blurring start point KPS, with the length of the figure ZK beginning at the start point ps and ending at the end point pe, for example, a mean value of the length of the f irst contour and the length of the second con-t our. being set to 1 Similarly, the blurring end length data is a value comprised within a range of f rom 0 to 1 obtained on normalizing the length from the starting point ps as far as the blurring end point PE, with the length of the figure being set to 1. This simp] if ies the blurring start length data and the blurring end length data. of course, the blurring start length data and the blurring end length data may be set so as to be equal to the actual length f rom the start paint ps as f ar as the blurring start point KP and the actual length f rom the start point ps as far as the blurring end point KPE.
With the blurred region data, the position within the f igure of the blurred area may be represented solely by the blurring start length data and the blurring end data, that is two one- c dimensionally arrayed data. In other words, with such blurred area data, the blurred area position may be specified by simplified data. In addition, if the blurring start data and t.he blurring end data are normalized based upon the figure length, data specifying the position of the blurred area may be simplified further.
The blurred area data is the data set for corresponding to a sole blurred area, so that, if there exist plural blurred areas in a figure, Plural blurred areas may be set in each blurred area.
Fig.33 shows an illustrative example of the above-described blurred area data. The blurred area data shown in Fig.33 is such data in which, for a figure having two blurred areas KS, eight blurring trajectories SK1, SK2, BK3, SK4, SK5, BK6, BK7 and BK8 are provided, with the blurred area data being normalized with the figure length set to 1. It is noted that the blurred area data shown in Fig.33 are simplified for explanation and that, for defining the blurred area KS more precisely, the number of the drawing traJectories, tha-. is the numbers of the blurring start length data and the blurring end length data, are set to larger values, while the numbers of the effective digits of the blurring start length data and the blurring end length data are also set to larger.values.
Refe'ring to Fig.34, the blurred area KS represented by the blurred area data shown in Fig.33 is comprised of a first bluIrred area KS1 surrounded by a blurring start paint KPi and a blurring end paint KP2 on the drawing trajectory BK2, a blurring start point. KP3 and a blurring end point KP4 an the drawing trajectory BK3, a blurring start point KP5 and a blurring end point KP6 on the drawing trajectory SK4, a blurring start paint KP7 and a blurring end point KP8 on the drawing trajectory BK5 and a blurring start point KP9 and a blurring end point KP10 on the drawing trajectory BK6, and a second blurred area K52 surrounded by a blurring start point KP11 and a blurring end point KP12 on the drawing trajectory BK4, a blurring start point KP13 and a blurring end point KP14 on the drawing trajectory BKS, a blurring start point KP15 and a blurring end point KP16 on the drawing trajectory BK6, a blurring start point KP17 and a blurring end point KP18 on-the drawing trajectory BK7 and a blurring start point KP19 and a blurring end point KP20 on the drawing trajectory BK8.
If the f igure length is set to 1, the lengths on the drawing trajectories E3K2, BK3, BK4, BK5 and SK6 from the start point as far as the blurring start points KPi, KP3, KPS, KP7 and KP9 are 0. 12, 0.31, 0. 25, 0. 14 and 0. 20, respect ively, whi le the lengths from the starting point as far as the blurring end points KPI, KP3, KP5, KP7 and KP9 are 0. 51, 0. 60, 0. 69, 0. 42 and 0. 55, respectively. On the other hand, the lengths on the drawing trajectori-es SK4, BKS, BK6, BK7 and BK8 from the start poin! as far as the blurring start points KPil, KP13, KPIS, KP17 and KP19 67 are 0.71. 0.84, 0.82, 0.78 and 0.73, respectively, while the lengths from the starting point as far as the blurring end points KP12, KP14, KP16, KP18 and KP20 are 0.92 0. 93, 0.91, 0.95 a-nd 0.94, respectively.
(4-1-3) Blurring Parameters The blurring parameters are parameters for forming blurred area data, and are comprised of a parameter cl employed for setting the blurring starting length data for the blurred area data for specifying the blurring start position along the length of the figure, a parameter c2 employed for setting the blurring end length data for the blurred area data for specifying the blurring end position along the length of the f igure, a parameter c3 specifying the upper end of the blurring along the width of the f igure, a parameter c, specifying the lower end of the blurring along the width of the f igure, a parameter c5 specifying the extent of variations in the blurring start position along the length of the figure, a parameter c6 specifying the extent of variations in the blurring end position along the length of the figure, and a pa.ra;-,,el:e, - c, specifying the density 0-1 the blurring along the width of the figure. These parameters cl to C, specif y the positions in the f igure of the blurred area KS, as wi 11 be explained subsequently, For these parameters c, to c,, specified values are set _for specif yin"gthe positions of the blurred area KS in the figure ZK, as shown for example in Fig,36.
68 That is, values of from 0 to 1, obtained on normalization with the figure length of 1 for representing the length from the start paint P-S as far as the blurring start position, are Set f-or the parameter c,. while values of from 0 to 1, obtained on normalization with the figure length of 1 for representing the length from the start point ps as far as the burring end point, are set for the parameter C2.
and cz is naturally c 1 5 c2.
The parameters c,, c2 may be set in a different manner provided the blurring start position or the blurring end position may thereby be known. For example, the parameter cl may be set for representing the mean position of the blurring start positions KPS on the drawing trajectories BK while the parameter c, may be set for representing the mean position of the blurring end positions KPE on the drawing trajectories BK. The parameters Cl, c2 may be of actual lengths instead of being normalized with the drawing length set to 1.
The parameters c,, c, are set to values of f rom 0 to 1 corresponding to the numICers of zhe drawing trajectories SK delimiting the upper and lower ends of the blurring normalized by the number m of the drawing trajectories BK. That is, of the m drawing trajectories BK, the (C3 x rn) 'th drawing trajectory represents-the upper end of the blurring area KS, while the_(C4 X m) 'th cfr-awing trajectory SK represents the lower end of the blurring area KS. These parameters c,, c4 del imit the blurring The relation of the parameters c, 69 range along the width of the figure ZK.
The values of these parameters c 31 C4 are set so that 1 C3 c41 will be larger and smaller for broader and narrower widths-of the blurred area KS, respectively. If the blurred area KS is offset along the width of the figure ZK in one or the other direction. the values of the parameters C3, c4 are both set to values closer to 0 or 1.
The parameters c3, c4 may be set in a different manner provided the upper and lower positions along the width of the figure may thereby be known. For example, the parameters C3 and C4 may be set so as to be equal to the length f ram one contour of the f igure ZK to the upper end of the blurred area and to the length from the other contour of the figure ZK to the lower end of the blurred area, respectively. Although the parameters C3, C4 are normalized with the number m of the drawing trajectories BK, the actual number of the drawing trajectories BK my naturally be employed without normalization.
The parameter C5 is set to a value within a range of f ram 0 to 0.5 X (C2 C,) flor specifying the extent, of varia-!"icns of the blurring start position, normalized with the figure length being set to 1. Similarly, the parameter c6 is set to a value within a range of f ram 0 to 0.5 x (cZ c,) for specifying the extent of variations -of the blurring end position, normalized with the figure len-9th being set to 1.
The parameters cS, c, may be set in a different manner C3 provided variations in the blurring start position or the blurring end position may thereby be known. Although the parameters CS, c6 are normalized with the figure length being set. to 1, the actual length may naturally be employed without normalization.
The parameter c7 is set to a value within the range of f rom o to 1 representing the blurring density normalized so that the absence of the blurring in the range defined by the parameters and C4 and t he p resence of t he bl u r r i ng i n t he ent i re range denote 0 and 1, respectively. If, for example, the parameter c, is 0.4 and the number of the drawing trajectories BK traversing the area def ined by the parameters cP C, is k, the (0.4_ x k) drawing trajectories and the ((1 - 0.4) x k) drawing trajectories among the drawing trajectories represent the blurred portion, that is the portion where there is no ink, and the nonblurred portion, that is portion where there is ink, respectively.
The parameter c, may be set. in any different manner provided the blurring density along the f igure width may thereby be known. For example, the parameter c7 may be set to an actual number of the drawing trajectories BK present in the blurred portion.
It is noted that, when a figure is actually drawn with a writing brush on the paper, such blurring may be produced in which the-extent of blurring is large, such that the amount of the ink present in the blurred portion is small. This sort of 71 0 blurring is hereinafter referred to as intense blurring. In the case of the intense blurring, variations in the blurring start position and-in the blurring end position tends to be diminished. Conversely, in the case of the blurring shown in Fig.38, the extent of blurring is small, such that the amount of the ink present in the blurred portion is large. This sort of blurring is referred to hereinafter as linear blurring. In the case of the linear blurring. variations in the blurring start position and in the blurring end position tends te be increased.
Thus. for representing the intense blurring, a larger value on the order of 0.8 to 1.0 is set for the parameter c7, while a smaller value on the order of 0.4 to 0.7 is set f or t he parameters cS and c6. On the other hand, for representing the linear blurring, a smaller value on the order of 0.4 to 0.7 is set for the parameter C7, while a larger value on the order of 1/6 to 1/8 is set for the parameters cS and c6.
These blurring parameters are set from cne blurred area to another, such that. if there are plural blurred areas, plural blurring parameters are set for coping with the respective blurred areas. Blurred area data are produced for coping with the respective blurred areas based upon the blurring parameters set from one blurred area to another.
(4-1-4) Shape Data The Mape data is data specifying the figure shape and is comprised of length data, width data, curvature data and curved 72 point data, as shown for example in Fig.39. The values of these data are calculated based upon the full contour data complemented so that the number of the coordinate points of the first contour data and the number of the coordinate points of the second contour data will be equal to each other.
The length data is the data specifying the length from the start Point up to the end point of a figure and is comprised of a length L, as shown in Fig.39. This length L is equivalent to the above-mentioned figure length. For this length L, a mean value of the length of the first contour found from the first contour data and the length of the second contour found from the second contour data is set. Specifically, the length o-f the first contour is a sum of the sequentially calculated values of the distances 'between neighboring coordinate points A j and Ap., of the f irst contour data, while the length of the second contour is a sum of the sequentially calculated values of the distances between neighboring coordinate points Bj and Bj+1 of the second contour data. That is, if the X-coordinate and the Y-coordinate of a coordinate point A, are XAj and YA1, respectively, the Xcoordinate and the Ycoordinate of a coordinate point Ai,, are XM+.1 and YAifl respectively, the X-coordinate and the Ycoordinate of a coordinate point Bi are X,i and Y,j, the X-coordinate and the Ycoordinate- of a coordinate point Bi+.1 a r e X91fl and Y3i+l, respectiv-!1y, the length L is found f rom the equation (4-1 ..(4-1) 73 0 (X3; - X3;. 2 ya. - yA. 1) 2 L X (XAi XA;. 1) 2 YA; - YI.l') 7 E 2 Z=1 lO The wid - th data is data specifying the figure width and is comp r i sed of w wi dt hs Wi ( i = 1, 2. 3,..., w) between the coordinate points Aj and the coordinate points Bi, as set for each of the coordinate points of the first or second contour data, where M is the number of the first or second coordinate points, as shown in Fig.39.
The curvature data is data specifying the bend of a figure and is comprised of w curvatures ri (i = 1, 2, 3..... w), a s set for each coordinate point of the first or second contour data. Specifically, for finding the curvature rj, a coordinate point Ci-, of a mid point between the coordinate point Aj-, and the coordinate Point Bi-,, a coordinate point Cj of a mid point between the coordinate point Ai and the coordinate point Bl. and a coordinate point Ci,, of a mid point between the coordinate point Aii and the coordinate point Bj+1 are found, and between a line interconnecting the coordinate point coordinate point Cj and a line interconnecting the point Cj and the coordinate point Ci., is calculated, ei being a curvature ri for the coordinate point A, or the coordinate point Bi. The angular range possible for the curvature. r'i is from 0 to iso, and the smaller or the largerthe angle, the larger and the smaller is the bend of the figure, respectively. It is noted that the curvature rl for the starting 74 an angle el Ci-, and the coordinate this angle point A, or B, and the curvature rw for the end point AW or BW is pre-set to 180 for specifying the bend-free state.
The curved or bend point data is the data for bend points where the curvature ri indicates a bend exceeding a pre-set angle. The curved point data is made up of the bend point position r_ p and the number of bend points r_ n. The bend point position r_p specifies the number i of the curvature ri exceeding a pe-set value. such as 115', while the number of bend points r n is the number of such curvatures ri. It is also possible to set only the curved point position r_ p as the curvature data and to f ind the number of curved points r_ n f rom the number of set bend Point Positions r_ p.
(41-5) Picture Data The pictUre data is data set f rom pixel to pixel for displaying a picture on the display device 5, as explained previously. The picture data for a given pixel is comprised of ink data specifying the state of the ink in the pixel and luminance data specifying the luminance of the pixel, as shown in Fig. 41. The ink data is comprised of ink/no ink data. specifying the presence or absence of the ink in the pixel, ink density data specifying the ink density in the pixel and the ink quantity data specifying the ink quantity in the pixel.
(4-1-6) Wri-ting Brush Data The w-riting brush data is data simulating a writing brush and is comprised of the ink quantity 10.,, (k = 1, 2, 3, -.., m) and the ink density ID k (k = 1, 2, 3,..---m). with the ink quantity and ink density being set in association with M drawing trajectories, as shown in Fig.42. Based upon the ink quantity 10k and the ink density Wk. the above-mentioned ink quantity data and the ink density data for the picture data are set, as will be explained subsequently.
(4-2) Construction of the Blurring Processor An illustrative construction of the blurring processor 40 is explained.
Referring to Fig.43, the blurring processor 40 includes a writing brush data setting unit 41 for setting writing brush data. and a data readout unit 42 for reading full-contour.data, such as fonts, formed by the font forming processor 20, from the external storage device 4 or the memory 3. The blurring processor also includes a blurring parameter setting unit 43 for setting blurring parameters based upon the full contour data supplied from the data readout unit 42 and a blurred area data forming unit 44 fo.forming blurred area data based upon the blurring parameters set by the blurring parameter set-Ling unit 43. The blurring processor similarly includes an ink data setting unit 45 for setting ink data in the pixels in a figure based upon the full contour data supplied from the data readout unit 42, writing brush data set in the writing brush data setting unit 41 and blurr-ed area data formed by the blurred area data forming unit 44, and a luminance data setting unit 46 f or sett i ng 76 0 luminance data in the pixel for which ink data has been set in the ink data setting unit 45.
The luminance data as set by the luminance data setting unit 46 is supplied to the display device 5 and the picture corresponding to the luminance data is displayed on the display device 5 for displaying a blurred figure on the display device 5.
(4-2-1) writing Brush Data Setting Unit The writing brush data setting, unit 41 includes a writing brush selecting unit 411 for selecting writing brush data from plural pre- formed writing brush data, and a writing brush data forming unit 412 for forming writing brush data based upon input data from the input device 1, as shown for example in Fig.44. The writing brush data selected by the writing brush data selecting unit 411 or the 'writing brush data formed by the writing brush data forming unit 412 is supplied to the ink data setting unit 45.
The writing brush selecting unit 411 reads out plural writing brush data from the rremory 3 or the external storage device 4 and selects one of plural writing brush data responsive to a signal entered by the user responsive to the user actuation. Specifically, if illustrative figures to be drawn with the respective writing brushes are displayed on the display device 5, and tI.FTe user selects writing brush data by selecting the desired illustrative figures, from among the plural illustrative 77 figures displayed on the display device 5, with the aid of the input device, such as a mouse. the writing brush selecting unit 411 selects writing brush data based upon a signal f rom the input device 1 corresponding to the user actuation. The writing brush data selecting unit 411 transmits the selected writing brush data to the ink data setting unit 45. The result is that the writing brush data required for the desired figure drawing may be easily selected by the user.
The writing brush forming unit 412 forms writing brush data based upon input data from the input device 1. Specifically, the writing brush forming unit 412 calculates the ink quantity at each point on the writing brush based upon the ink quantity values for the respective points on the writing brush, as entered by the input "device 1, as shown for example in Fig.45. The writing brush forming unit 412 calculates the ink density at each point on the writing brush based upon the ink density values for the respective points on the writing brush, as shown for example in Fig.46. Data entry by the input device 1 may be simplified in this manner by entering only the approxirnnate distribution olthe ink density or quantity contained in the writing brush and by forming new writing brush data at the writing brush input device.
For further simplification, it is. also possible for-the writing brsh forming unit 412 to calculate the ink quantity at each point n the writing brush based upon the entire ink quantity 78 in the writing brush as entered by the input device 1. as shown for example in Fig.47. It is similarly possible for the writing brush forming unit 412 to calculate the ink density at each poi-nt on the writing brush based upon the entire ink density in the writing brush as entered by the input device 1, as shown for example in Fig.48.
(4-2-2) Data readout Unit The data readout unit 42 includes a plural contour data reading unit 421 for reading one or more full-contour data from the input device 1. memory 3 or the external storage device 4, a contour data selecting unit 422 for selecting the full-contour data corresponding to the blurred picture from the full-co.ntour data f rom the plural contour data reading unit 421 in accordance with a preset rule, and a contour data complementing unit 423 for selecting one or more full-contour data suppl ied f rom the plural contour data reading unit 421, as shown for example in Fig.49.
The plural contour data reading unit 421 reads one or more full-conlour data from trie input device 1, memory 3 or' the external storage device 4. Specifically, one or more full-contour data representing a blurredcharacter or symbol having at least one stroke, that is the font prepared by the above-rlentioned font forming un----it 20.
If orre or more full-contour data read by the plural contour data reading unit 421 represents a character or a sy.-,,bol, that 79 0 is if the respective f ul]-contour data correspond to the strokes constituting the character or the symbol the contour data selecting unit 422 selects the full contour data correspondi.ng to the blurred picture, f rom among the f ull-contour data, in accordance with the preset rule. The f igure represented by the full contour data selected by the contour data selecting unit 422 is to be an object of the blurred picture, while the f igure represented by the non-selected full contour data is not to be an object of the blurred picture. The contour data selecting unit 422 sends the information as to which full-contour data is to be an object of the blurred picture to the blurring parameter setting unit 43 and to the ink data setting unit 45.
The reason is that, if a character is actually written on the paper with a writing brush, blurring is not necessarily produced in all of the strikes constituting the character such that a blurred stroke and a nonblurred stroke usually cc-exist in the character. There exists a certain statistic rule between the burred st roke and the non-blurred stroke, as will be explained subsequently. The contour data selecting unit 422 selects the fullcontour data constituting the object of the blurred picture in accordance with the pre-set rule.
Specifically, this pre-set rule states that the ratio of the numbe r of the se 1 ect ed f u 11 cont ou r dat a t o t he numbe r of _t he entire ful-icontour data be not more than a pre-set value, such as one-third. That is, the contour data selecting unit 422 first selects, from all of the full-contour data supplied to the contour data selecting unit 422, the f ul 1 contour data having the longest stroe represented by the f ul 1 contour data. The contour data selecting unit 422 also selects, from among he full contour data excluding the full contour data directly before and after the previously selected longest full-contour data, the f ul 1concur data having a longer stroke represented by the fullcontour data. The contour data selecting unit 422 iterates he above operation until the number of the selected full contour data reaches one-third of the entire fullcontour data. The pre set rule may also be such a rule that full-contour data be selected from one or more full contour data supplied to the contour data selecting unit 422 so that the ratio of the selected fullcontour data to the total number of the full-contour data is not more than the pre-set ratio, such as one- third, and only one full-contour data be selected at random from the remaining full-contour data.
By selecting the full-contour data as an object of the blurred picture from one or more ofl. the full-contour data n accordance wit the pre-set rule and by drawing the character so that blurring is produced in the stroke represented by the selected full-contour data but is not produced in the stroke represented by the non-selected full-contour data, the character may be drawn to a more beautiful appearance.
However, if one or more full-contour data read by the plural 81 full contour data reading unit 421 represents a f igure other than the character or the symbol, the contour data selecting unit 422 selects all of the full-contour data read by the plural fu.11 contour data reading unit 421 as the full contour data which is to be the object of the blurred picture.
On the other hand, the contour data complementing unit 423 includes a contour data supplementation unit 424 f or supplementing cordate points to the first contour data and the second contour data of the f ul 1 contour data suppl ied f rom plural contour data reading unit 421 and a contour the data adjustment unit 425 for coinciding the number of the coordinate points of the first contour data and that of the coordinate points of the second contour data of the full contour data supplied from the plural contour data reading unit 421, as shown for example in Fig.49.
The contour data complementing unit 423 supplements the coordinate points to the first contour data and the second contour data by the contour data supplementation unit 424 for producing new coordinate data having a smal ler number between the coordinate Points and further supplements coordinate points to the new full coordinate points by the contour data adjustment unit 425 for forming new full-contour data in which the number of the coordinate points of the first contour data is equal_to that of t" coordinate points of the second coordinate data.
Specifically, if the distance between the neighboring 82 coordinate points Ai and Ai., of the f irst contour data is larger than a pre-set value, the contour data supplementation unit 424 supplements coordinate points between the neighboring coordinate points Aj and Aj+1 for forming new f irst contour data so that the distance between the neighboring coordinate points will be smaller than a pre-set value. On the other hand, if the distance between the neighboring coordinate points Bi and Bi+, of the second contour data is larger than a preset value, the contour data supplementation unit 424 supplements coordinate points between the neighboring coordinate points Bi and Bi., for forming new second contour data so t hat the distance between the neighboring coordinate points will be smaller than a pre-set value. The Pre-set value is selected so that the first and second contours represented by the new first and second contour data will appear smooth.
The contour data supplementation unit 424 supplements coordinate points so that the f i rst contour produ--j-:d on sequentially interconnecting the coordinate points Ai and the second contour produced on sequentially interconnecting the coordinate points Bj using the spline curve or the Bez ier curve will become smooth. Specifically, the contour data supplementation unit 424 sets the direction of a tangential line at the coor-dinate point A, at the starting point as the direc-icn proceeding from the coordinate point AI towards the coordinate point A, for setting a spline curve traversing the coordinate 83 points Ai, and supplements the coordinate points on the spline curve to the first contour data. The contour data supplementation unit 424 also sets the direction of a tangential line at t.he coordinate Paint B, at the starting point as the direction proceeding from the coordinate point B1 towards the coordinate point B2 for setting a spline curve traversing the coordinate points Bi, and supplements the coordinate points on the spline curve to the second contour data.
By supplementing the coordinate points at the cordate data supplementation nit 424, the first and second contours become smooth thus enabling a figure to be drawn with more beautiful appearance.
The contour data supplementation unit 424 sends to the contour data adjustment unit 425 full-contour data in which the distance between the coordinate points becomes smaller than the pre-set value and the first and second contours have become smooth.
If, in the full-contour data supplied from the contour data supplementation unit 424, the numCer of the coordinate points oll the first contour data and the number of the coordinate points of the second contour data are different from each other, the contour data adjustment unit 425 supplements the coordinate points to the first contour data or the second contour dataso that the n-umber of the coordinate points of the first and second contour data will become equal to each other, in order to from 84 new first and second contour data.
Specifically, the contour data adjustment unit 425 selects two neighboring coordinate points having a broader distance between the coordinate points, beginning from one of the first and second contour data having the smaller number of the coordinate Points. If the distances between the coordinate points are all the same, the contour data adjustment unit 425 selects two coordinate points at random. The contour data adjustment unit 425 supplements a coordinate point which is on a straight line, Bezier curve or on a spline curve interconnecting these two coordinate points and which is equally spaced apart from these coordinate points. This enables the coordinate point to be supplemented to render the first and second contours smooth and to reduce variations in the distance between the coordinate points.
Thus the contour data supplementation unit 423 supplements the new coordinate point to the full contour data supplied from the plural contour data reading unit 421 to form new full-con-Lour data. The full- contour data supplementation unit 423 supplies the new full contour data to the blurring parameter setting unit 43 and to the ink data set-Ling unit 45. (4-2-3) Blurring parameter Setting Unit Referring to Fig.50, the blurring parameter setting unit 43 includes A blurring parameter selecting unit 431 for reading plural pre-formed blurring parameters from the memory 3 or the external storage device 4 for selecting blurring Parameters f rom the plural blurring parameters responsive to the signal f rom the input device- 1 by user actuation, and a first blurring paramet.er calculating unit 432 for calculating blurring parameters based upon the writing brush carrying speed and writing brush pressure distribution entered from the input device 1 as the writing brush is carried for figure drawing from the starting point to the end point. The blurring parameter setting unit also includes a second blurring Parameter calculating unit 433 for calculating the blurring parameters based upon the shape of the figure represented by the full contour data supplied from the data readout unit 42, and a blurring parameter input device 434 for forming the blurring parameters from the parameters c, to C7 entered via the input device 1.
The blurring parameter selection nit 431 sends the selected blurring parameters to the blurred area data forming unit 44, while the first blurring parameter calculating unit 432 sends the calculated blurring parameters to the blurred area forming unit 44. The second blurring paramezer call cul at ing unit 432 sends the calculated blurring parameters to the blurred area forming unit 44. The blurring parameter input device 434 sends the formed blurring parameters to the blurred area forming unit 44.
The blurring parameter selecting unit 431 selects blurring parametersfrom the plural pre-formed blurring parameters having different combinations of the values of the parameters c., to C7 86 supplied from the memory c,necifically 3 or the external storage device 4. typical b 1 u r r i n 9 examples obtained using the blurring parameters are displayed on the display device 5. -If the user selects, from the plural blurring examples displayed on the display device 5, the desired blurring example by the input device, such as a mouse, for thereby selecting the blurring parameters, the blurring parameter selection unit 431 selects the blurring Parameters based upon the signal from the input device 1 responsive to the user actuation. The blurring parameter selection unit 431 sends the selected blurring parameters to the blurred area data forming unit 44. The result is that the user may easily select the blurring parameters required for drawing the desired figure.
The first blurring parameter calculating unit 432 calculates the parameters cl to c,, based upon the speed distribution data and the writing brush distribution data entered from the input device 1. and sends the blurring parameters comprised of these parameters c, to C7 to the blurred area data forming unit 44.
The speed distribution data is the data specifying the distribution of the writing brush carrying speed during figure drawing from the beginning paint to the terminal point, and is comprised of 1 _q speeds v, (i = 1, 2, 3,..., 9) specifying-the writing brCish carrying speed at the respective a positions within the figure, as shown in Fig.51. The writing brush distribution 87 data is the data specifying the distribution of the writing brush pressure during figure drawing from the beginning point to the terminal point, and is comprised of _q pressures pi (i = 1, 2, -3,... 1 g) specifying the writing brush pressure at the respective q positions within the figure, as shown in Fig.51.
For the input device 1 for entering the speed distribution data and the writing brush pressure distribution data, a tablet is employed. That is, the user enters the pen movement speed with the tablet as the speed distribution data, while entering the pen pressure as the writing brush distribution data. This makes it possible to enter the speed distribution data and the writing brush distribution data more easily.
The second blurring parameter calculating unit 433 calculates the parameters c, to C7, based upon the shape of the figure represented by the full-cont(our data supplied from the data readout unit 42 as later explained, and sends blurring parameters comprised of the parameters cl to c, to the blurred area data forming unit 44.
The blurred parameter input unit. 434 directly sends the parameters cl to c7 entered from the input device 1 as blurring parameters to the blurred area data forming unit. 44.
The blurring parameter setting unit 43 decides, based upon the signalfrom the input device 1 responsive to the operation by the usd-r, which of the blurring paramet-er selecting unit 431, first blurring parameter calculating unit 432, second blurring 88 parameter calculating unit 433 and the blurring parameter input device 434 is to be used in setting the blurring parameters, If the user desires to set the blurring parameters while viewing the blurring manner, he or she selects the blurring parameter selecting unit 431. If the user desires to set the blurring parameters based upon the actual writing brush carrying speed or actual writing pressure, he or she selects the first blurring parameter calculating unit 432. If the user desires to set the blurring parameters for generating usual blurring, he or she selects the first blurring parameter calculating unit 432. Finally, if the user desires to set the blurring parameters freely, he or she selects the blurring parameter input device 434.
(4-2-4) Blurred Area Data Firming Unit The blurred area data forming unit 44 forms the blurred area data based upon the blurring parameters supplied from the blurring parameter setting unit 43, as will be explained subsequently. The blurred area data forming unit 44 sends the formed blurred area data to the ink data setting unit 45.
(4-2-5) Ink data Setting Unit Referring to Fig.52, the ink data setting unit 45 includes a first Pixel detection unit 451 for detecting the pixel based upon full contour data supplied from the data reading unit 42 and a fir"gt picture data setting unit 452 for setting picture data on the pixel detected by the first pixel setting unit 451.
89 The ink data setting unit 45 also includes a second Pixel detection unit 453 for detecting the pixel based upon the pixel detected by.the first pixel detection unit 451 and a second picture data setting unit 454 for setting picture data on the pixel detected by the second pixel setting unit 451.
The first pixel detection unit 451 detects, based upon the full-contour data supplied from the data reading unit 42, the pixels on a line segment 1 i interconnecting the coordinate point Ai of the f i rst contour data and the coordinate point Bi of the second contour data. One or plural pixels, usually plural pixels, are detected i n t hi s manne r. The range of J i s f rom 1 to w. The pixel detected for i = 1 is that on the starting point, -while that detected for i = w is that on the end point.
The first picture data setting unit 452 includes an ink/no ink setting unit 455, a first ink density setting unit 456 and a first ink quantity setting unit 457. The first picture data setting unit 452 sets ink data in the pixel detected by the f i rst pixel detection unit 451 based upon the writing brush data supplied f rom the writing brush data setting unit 41, fullcontour data supplied f rom the data reading unit 42 and the blurred area data suppI ied f rom the blurred area data f orming unit 44 and sends the ink data to the second picture data setting unit 454 and the luminance data setting unit 46.
The i-nk/no ink data setting unit 455 judges, based upon the full-contour data and the blurred area data, whether or not the pixel detected by the first pixel detection unit 451 is within the blurred area KS. If the pixel is within the blurred area KS, the ink/no ink setting unit 455 sets the ink/no ink data for specifying the state of ink absence in the pixel. If the pixel is not within the blurred area KS, the ink/no ink setting unit 455 sets the ink/no ink data for specifying the state of ink presence in the pixel. If there are plural pixels detected by the first pixel detection unit 451, the ink/no ink setting unit 455 naturally judges whether or not the pixels are separately in the blurred area for setting the ink/no ink data.
If the blurring start length and the blurring end length indicated by the blurring start length data and by the blu-rring end length data of the blurred area data associated with the figure drawing trajectory traversing the pixel are denoted as Ks and Ke, respectively, and the figure length from the starting point as far as the line segment li is Li, the ink presence/ink absence data setting unit 455 judges the pixel as being within and outside the blurred area KS for Ks < Li < Ke and for Li5 Ps cr Ke 5 Li, respectively. This judgment, however, is given for each blurred area KS if plural blurred area data are set. Th a t is, if the pixel being judged is present in the blurred area KS represented by any of the blurred area data, the pixel is judged to be within the blurred area KS. Conversely, if the pixel being judged is -present outside the blurred area KS represented by al 1 of the blurred area data, the pixel is judged to be outside the 91 blurred area KS.
On the other hand, the first ink density setting unit 456 sets ink density in the pixel detected by the f i rst pixel detection unit 451, based upon writing brush data associated with the figure drawing trajectory traversing the pixel detected by the f irst pixel detection unit 451, so that the ink density wil 1 be higher the shorter the length of the f igure f rom the starting point as far as the line segment li. The first ink density setting unit 456 sends the ink density data to the second picture data setting unit 454 and to the luminance data setting unit 46.
Thefirst inkquantity setting unit 457 sets the ink density data in the pixel detected by the first pixel detection unit 451, based upon writing brush data associated with the figure drawing trajectory traversing the pixel detected by the first pixel detection unit 451, so that the ink quantity will be larger the shorter the length of the figure from the starting point as far as the line segment li. The first, ink quantity setting unit 457 sends the ink quantity data to the second picture data setting unit 454 and to the luminance data setting unit 46.
However, if the ink presence/ink absence data is set for the pixel for specifying the ink absent data, the first ink density setting unit 456 sets the ink density data to 0, while the first ink quanti_ty setting unit 457 sets the ink quantity data to 0. If there are plural pixels detected by the first pixel detection unit 451, the first ink density setting unit 456 and the first.
92 ink quantity setting unit 457 natural 1 Y calculate the ink density and the ink quantity on the pixel basis for setting the ink density data and the ink quantity data, respectively.
The second pixel detection unit 453 detects a pixel lying between the pixel on the line segment li, as detected by the first pixel detection unit 451, and a pixel on a line segment 1 iff neighboring to the line segment],, as detected by the first pixel detecting unit 451. Usually, plural pixels are detected by the second pixel detection unit 453. If the pixel on the line segment Ii and the pixel on the line segment li,., neighboring to the line segment li are adjacent to each other, zero pixels are detected by the second pixel detection nit 453.
The second picture data setting unit 454 includes a second ink density setting unit 458 for setting ink density data in the pixel detected by the second pixel detection unit 453 and a second ink quantity setting unit 459 for setting ink quantity data in the pixel detected by the second pixel detection unit 453.
The second ink density setting unit 458 calculates, based upon ink density data set in the pixel on the 1 ine segment 11 supplied from the first picture data setting unit 452, distance between the pixel on the line segment Ii and the pixel detected by the second pixel detection unit 453, ink density data set in the pixel-on the line segment ii,-, supplied from the first picture data setting unit 452, and the distance between the pixel on the 93 line segment 1j., and the pixel detected by the second pixel detection unit 453, the ink density in the pixel detected by the second Pixel. detection unit 453, sets ink density data an the pixel detected by the second pixel detection unit 453 and sends the ink density data to the luminance data setting unit 46. 1 The second ink quantity setting unit 459 calculates, based upon ink quantity data set in the pixel on the line segment 1, supplied from the first picture data set-Ling unit 452, distance between the pixel on the line segment Ii and the pixel detected by the second pixel detection unit 453, ink quantity data set in the pixel on the line segment li,l supplied from the first picture data setting unit 452, and the distance between the pixel on the line segment 1j+1 and the pixel detected by the second pixel detection unit 453, the ink quantity in the pixel detected by the second pixel detection unit 453, sets ink quantity data on the pixel detected by the second pixel detection unit 453 and sends the ink quantity data to the luminance data setting unit 46.
(4-2-6) Luminance Data Setting Unit The luminance data setting unit 46 calculates luminance in each pixel, based upon the paper fiber structure data formed by the above-described paper fiber structure data forming unit and picture dTta set in the first and second picture data set-ting units, and sets luminance data for each pixel.
The luminance data setting unit 46 calculates the luminance 94 data for each pixel, based upon the numbers of fibers M represented by the paper f iber structure data and the ink density and ink quantity data set in the first picture data setting unit 452 and in the second picture data setting unit 454, as will be explained subsequently. However, for pixels for which the picture data has not been set in the first picture data setting unit 452 or in the second picture data setting unit 454. default data, such data specifying pure white or data specifying pure black, are set.
The luminance data setting unit 46 calculates luminance using only data of the numbers of fibers M, among data constituting the paper fiber structure data. Thus it sufjices for the paper fiber structure data referred to here to have at least the numbers of fibers M, while it is unnecessary for the data to have e.g., the number of capillary tubes.
(4-3) Blurring Processing An illustrative operation of the blurring processor 40 is explained.
If the blurring selection unit 431 or the first blurring parameter calculating unit 432 is used for setting blurring parameters, the blurring processor 40 performs blurring processing by the following steps (a) to (d):
(a) The step of setting writing brush data by the writing b-rush data settt-ng unit 41; (b) the step of setting the blurring parameters by the blurring parameter setting unit 43 using the blurring parameter selection unit 431 or the first blurring parameter calculating unit 432 and subsequently forming blurred area data by the blurred area data forming unit 44 based upon the blurring parameters set by the blurring Parameter setting unit 43; (c) reading one or more full-contour data by the data reading unit 42 from e.g., the input device 1, memory 3. external storage device 4 or the font forming processor 20 and selecting the full contour data as an object of blurred drawing by the contour data selecting unit 422; and (d) setting ink data in pixels in a f igure by the ink data setting unit 45 based upon the full-contour data suppliedfrom the data reading unit 42, writing brush data set by the writing brush setting unit 41 and the blurred area data formed by the blurred area data forming unit 4 and subsequently setting luminance data in he pixels for which the ink data has been set by the ink data setting unit 45.
The steps (a) to (c) may be carried out in any sequence provided that these steps are carried out prior to the step (d).
If the second blurring parameter calculating unit 433 is used for setting the blurring parameters, the blurring processor 40 performs blurred drawing by the following steps (e) to (h):
(e) The"step of setting writing brush data by the writing brush data setting unit 41; (f) reading out one or more full-contour data from the input 96 device 1, memory 3, external storage device 4 or the font forming processor 20 by the data readout unit 42 and selecting the fullcontour data -as the object of blurred drawing by the contour data selecting unit 422; (g) setting blurring parameters by the blurring parameter setting unit 43, using the second blurring parameter calculating unit 43, based upon the full-contour data selected as the object of blurred drawing by the blurring data selecting unit 422 and subsequently forming blurred area data by the blurred area data forming unit 44 based upon the blurring parameters set by the blurred parameter setting unit 43; and (h) setting ink data in the pixels in the f igure by the ink data setting unit 45 based upon the full-contour data supplied by the data reading unit 42, writing brush data set by the writing brush setting unit 41 and blurred area data formed by the blurred area data forming unit 44 and subsequently setting luminance data in the pixels in which the ink data has been set by the ink data setting unit 45.
The step (e) may be set at any stage provided it is performed before the step (h).
Of the blurring drawing processing steps, the setting of the blurring parameters by the blurring parameter setting unit 43, formation -_of the blurred area data by the blurred area data forming "u-nit 44, selection of the full-contour data by the contour data selecting unit 422 and the setting of the ink data 97 by the ink data setting unit 45, are explained in detail.
(4-3-1) Setting of Blurring Parameters by Blurring Parameter Setting UnitReferring to the flowchart of Figs.53 to 57, the illustrative operation of the blurring parameter setting unit 43 is now explained.
At step ST-4-1-1, a command as to which of the blurring parameter selection unit 431, first blurring parameter calculating unit 432 or the second blurring parameter calculating unit 433 is to be employed is entered f rom the input device 1 to the blurring parameter setting unit 43.
Then, at steps ST4-12, ST4-1-3, ST4-1-4 and ST4-1-5., the blurring parameter setting unit 43 judges, based upon the command entered at step ST4-1-1, which of the blurring parameter selection unit 431, f irst blurring parameter calculating unit 432, second blurring parameter calculating unit 433 or the blurring parameter input device 434 is to be employed. If the blurring parameter selecting unit 431 is to be used, the blurring parameter setting unit 43 transfers to step STS1741-21, whereas, if the first blurring parameter calculating unit 432 is to be used, the blurring parameter setting unit 43 transfers to step STS174-1-31. If the second blurring parameter calculating unit 433 is to be dsed, the blurring parameter setting unit 43 trans-fers to step S-TS1-4-1-41, whereas, if the blurring parameter input device 434 is to be used, the blurring parametersetting unit 43 98 transfers to step STS174-1-51.
Referring to Fig.54, at step ST4-1-21. to which the blurring parameter setting unit transfers if the blurring parameter is set using the blurring parameter selecting unit 431 ' a command on which of previously formed plural blurring parameters is to be used is entered from the input device 1 to the blurring parameter selecting unit 431. Specifically, by displaying examples of blurring produced with the use of each of the preformed plural blurring parameters an the display device 5, and by the user selecting a desi red one of the blurring examples displayed on the display device 5 using the input device 1, such as a mouse, a command as to which of the blurring parameters is to be employed is entered to the blurring parameter selection unit 431. Thus the user is able to easily command the blurring parameter required for generating the desired blurring.
Then, at step ST4-1-22, the blurring parameter selecting unit 431 selects one of the previously formed plural blurring parameters, based upon the command entered at step ST4-1-21, and sets the selected tlurring paramneter to co.-,iplet.e the processing.
At step ST4-1-31, to which the f irst blurring parameter calculating unit transfers if the blurring parameter is set using the first blurring parameter calculating unit 432, speed distributi.on data and writing brush pressure distribution data are entered f rom the input device, such as a tablet, to the f i rst blurring parameter calculating unit 432. If, however, the tablet 99 is used as the input device 1, input data tend to be f luctuated in an unexpected manner at the input time due to pen wobbling. Thus, if the speed distribution or writing brush pressure distribution be changed significantly within a short period, these changes are assumed to be ascribable to unexpected changes in the speed distribution or writing brush pressure distribution caused by pen wobbling at the input time and data correction is made based upon temporally previous or temporally subsequent speed distribution or writing brush distribution.
Then, a 'L step ST4-1-32, the first blurring parameter calculating unit 432 converts the speed distribution data and the writing brush pressure distribution data entered at step ST4:-1-31 into bi-level signals. The f i rst blurring parameter calculating unit 432 then transfers to step ST41-33. That is, the first blurring parameter calculating unit 432 processes the speed distribution data so that the speeds vi (i = 1, 2, 3, -., 9) at respective Points in the writing brush carrying direction higher and not higher than a pre-set value, with the writing brush carrying speed being then higher and lower, respectively, are set to 1 and 0 respectively. Similarly, the first blurring parameter calculating unit 432 processes the writing brush pressure distribution data so that the writing brush pressures pi (i = 1, 2, 3, 9) at respective points in the writing brush carrying d41rectior-higher and not higher than a pre-se- !,- value, with the writing brush pressure being then higher and lower, respectively, are set to 1 and 0, respectively.
Specifically, if fed with the speed distribution data and the writing brush distribution data yet to be converted into bilevel signals are entered from the input device 1, as shown for example in Fig.58, the first blurring parameter calculating unit 432 sets the pre-set value to e. g., 0.5, and forms speed distribution data and the writing brush pressure distribution data in the bilevel form, as shown in Fig.59. The speed distribution data and the writing brush distribution data shown in Figs. 59 and 60 are simplif ied for explanation and, in effect, the data yet to be converted into bi-level signals are more abundant in the number of effective digits or in the overall data volume.
At step ST4-1-33, the first blurring parameter calculating unit 432 forms blurring brush-carrying data baed upon the speed distribution data and the writing brush distribution data converted into the bi-level signals at step ST4-1-32. The first blurring Parameter calculating unit 432 then transfers to step ST4-1-34.
Specifically, the first blurring parameter calculating unit 432 forms blurring brush-carrying data made up of a blurring data ti (i = 1, 2, 3,. _ 9) specifying the state of blurring during brush carr--ying at respective _q positions within the figure.- The values of-the blurring data ti at these a points are in a range of f rom 0 to 1. The stronger and the weaker the r-..a.nr,,er of 101 blurring, the larger and the smaller is the value of the blurring data t j, respect i ve 1 y.
That isI if the speed vi is 0 and the writing brush pressure Pi is 0, the speed is low and the writing brush pressure is also low, so that blurring is assumed not to occur significantly. and hence the blurring data ti is set to a small value, such as 0.3. If the speed vi is 1 and the writing brush pressure pi is 0, the speed is high but the writing brush pressure is low, so that blurring is assumed to occur and hence the blurring data ti is set to the maximum value, that is 1. If the speed vi is 0 and the writing brush pressure pi is 1, the speed is low but the writing brush pressure is high, so that blurring is assumed not to occur and hence the blurring data ti is set to the minimum value, that is 0. If the speed vi is 1 and the writing brush pressure p, is 1, the speed is high and the writing brush pressure is also high, so that the blurring is assumed to occur to some extent and hence the blurring data ti is set to a larger value, such as 0.5.
As typical of the blurring bush-carrying data, the blurring brushcarrying data, formed based upon the speed distribution data and the writing brush pressure distribution data, shown in Fig.59, are shown in Fig.60.
At step ST-4-1-34, the first blurring parameter calculating unit 432 calculates the blurring parameters, based upon the blurring brushcarrying data formed at step ST4-1-33, and sets 102 the calculated data as the blurring parameters, to terminate the processing.
At step ST4-1-41 in Fig.56, to which the processing transfers when the blurring parameters are set using the second blurring Parameter calculating unit 433, the second blurring parameter calculating unit 433 forms shape data representing the shape of a figure represented by the full-contour data supplied from the contour data reading unit 42. The second blurring parameter calculating unit 433 then transfers to step ST41-142.
At step ST4-1-42, the second blurring parameter calculating unit 433 calculates the blurring parameters, based upon the shape parameters formed at step ST4-1-41, and sets the calculated parameters as the blurring parameters, to terminate the processing.
At step ST4-151, shown in Fig.57, to which the processing transfers when the blurring parameters are set using the blurring parameter input device 434, the parameters c, to c7 are entered from e.g., the input device 1 to the blurring parameter input device 434.
Then, at step ST4-1-52, the blurring parameter forming unit directly sets the parameters c, to c, entered at step ST4-151 as the blurring parameters to terminate the processing.
If plral blurred areas are resent in a figure, that is if plural blurred area dalla are set, plural sets of the blurring parameters are set in association with the respective blurred 103 area data.
(4-3-2) Formation of Blurred Area Data by Blurred Area Data Forming UnitAn illustrative operation of the blurred area data forming unit 44 is explained by referring to the flowchart of Fig.61.
At step ST42-1, the blurred area data forming unit 44 initializes the blurred area data and sets all blurring start length data and al blurring end length data to e.g., 1. The blurred area data forming unit 44 then transfers to step ST4-2-2.
At step ST4-2-2, blurred area data forming unit 44 refers to the value of the parameter C7 and transfers to steps ST4-2-3 and ST4-2-4 if the value of the parameter C7 'S 1 (YES) and is not 1 (NO), respectively.
At step ST4-2-3, the blurred area data forming unit 44 sets values to the blurring start length data and the blurring end length data, associated with all drawing trajectories positioned within the width-wise range as set by the parameters c3 and C4. That is, if the value of c, is 1, the width-wise range of blurring def ined by the parameters c. and c, represents the blurred area. The blurred area data forming unit 44 sets values on these blurring start and end length data to terminate the processing.
Spe.cif ical 1 y. the blurred area data forming unit 44 set---s at random a value ranging f rom c, to (cl + cS) as the blurring start length data, while setting at random a value ranging f rom c2 to 104 (C2 + c6) as the blurring end length data. Thus the blurring start length data and the blurring end length data, associated with the drawing trajectories located within the widthwise ran---ge of blurring as def ined by the parameters c3, C4. are f luctuated within the range of f rom c 1 to (cl + C5) and within the range of from C2 to (c, + c6), respectively.
At step ST4-2-4, the blurred area data forming unit 44 calculates how many of the drawing trajectories positioned within the width-wise blurring range as def ined by the parameters c3 and. The blurred area data formi ng uni t 44 t hen t rans fe rs to step ST4-2-5. That is. if the val ue of val ue of c, i s not 1, a] 1 of the wi dt h-wi se bl ur r i ng range defined by the parameters c3 and c, is not the blurred area. Specifically, the number n of the drawing trajectories associated with the blurred area is calculated as n = m x 1C3 C41 x cl, where m is the total number of the drawing trajectories.
At step ST4-2-5, the blurred area data forming unit 44 selects a! random one of the drawing traJectories located within the width-wise blurring range defined by the paramete rs c3 and C4, and sets values in the blurring start length data and the blurring end length data associated with the drawing trajectories. The blurred area data forming unit 44 then trans'lers-to step ST4-2-6.
Specifically, the blurred area data formina unit 44 sets at C4 are associated with the blurred area random a value ranging f rom cl to (c, + c5) for the blurring start length data, while setting at random a value ranging from c, to (c2 + c6) for the blurring end length data. Thus the blurring start length data and the blurring end length data, associated with the drawing trajectories located within the width-wise range of blurring as def ined by the parameters C3, c,, are f luctuated within the range of f rom c, to (c, + c5) and within the range of from c2 to (C2 + C60, respectively.
Meanwhile, if values other than initial values are set in the blurring start length data and the blurring end length data, associated with the selected drawing trajectories, that is if the processing corresponding to step ST4-25 is previously performed, the blurred area data forming unit 44 selects the next drawing trajectories, without updating the values of the blurring start length data and the blurring end length data associated with the preselected drawing trajectories.
At step ST4-2-6, the blurred area data forming unit 44 compares the number of times of the processing operations of the step ST4-2-5. thlat is the number of the blurring start' length data and the blurring end length data, for which values other than the initial values have been set, to the number n of the drawing trajectories associated with the blurred area, calculated at step ST4-2-4. If the result is NO, that is if the numbe,- of ti,-,ies of processing at step ST4-2-5 is less than the number n of the drawing trajectories calculated at step ST4-2-4, the blurred 106 area data forming unit 44 reverts to step ST4-2-5 to repeat the processing. If the result is YES, that is if the number of times of processing at step ST4-2-5 has reached the number n of the drawing trajectories calculated at step ST4-24. the processing comes to a close.
If there are plural blurred areas, that is if plural blurring parameters are set, plural blurred area data are set in association with the respective blurring parameters.
(4-3-3) Selection of FullContour Data by Contour Data Selection Unit The illustrative operation of the contour data selecting unit 422 is explained by referring to the flowchart of Fig.62.
At step ST43-1, the contour data selection unit 422 selects f rom all fullcontour data- the full-contour data having the longest length of the f igure represented by the full-contour data. The contour data selection unit 422 then transfers to step ST4-3-2.
At step ST4-3-2, the full-length data selecting unit 422 judges the number of the selected full-contour data with respect to the number of al 1 f ul]-contour. If the number of the selected f ul 1-contour data is not more than one-thi rd (YES), the f ul]length data selecting unit 422 transfers to step ST4-3-3. If the numbe r o.f t he se 1 ect ed f u 11 - cont ou r dat a exceeds onet h i r d (-NO) the full.-length data selecting unit 422 transfers to step ST4-3- 4.
107 0 At step ST4-3-3, the fulllength data selecting unit 422 selects one non- selected full-contour data, from among the fullcontour datwexcluding the full-contour data directly before and after the full-contour data selected at step ST4-3-1, which has the longest length of the figure represented by the full-contour data. The full-length data selecting unit 422 then reverts to step ST4-3-2 to repeat the processing.
On the other hand, the full-length data selecting unit 422 selects at random one full-contour data, from among the nonselected full-contour data, to terminate the processing.
(4-3-4) Setting of Ink Data by Ink Data Setting Unit The illustrative operation of the ink data setting unit 45 is explained by referring to the flowchart of Fig.63.
At step ST4-4-1, the first pixel detection unit 451 sets the value of the number j of times of processing f lag to 0. before transferring to step ST4-4-2.
At step ST4-4-2, the first pixel detection unit 451 increments the value of the number i of times of processing f lag by 1, before transferring to step ST44-3.
At step ST4-4-3, the first pixel detection unit 451 compares the number w of the coordinate points of the f i rst or second contour data to the value i of the number of times of processing flag. If i:5 w (YES), the processing transfers to step ST4-4-4. if i > W (NO), the processing transfers to step ST4-4-8.
At step ST4-4-4, the first pixel detection unit 451 detects 108 the pixels on a line segment Ii interconnecting the coordinate point Ai of the first contour data and the coordinate point Bi of the second contour data, where i denotes the value of the number of times of processing flag J.
Then. at step ST4-45, the ink presence/ink absence setting unit 455 refers to blurred area data in order to judge whether or not the Pixel detected at step ST4-4-4 is within the blurred area. If the pixel is found to be within the blurred area, the ink presence/ink absence data is set for indicating the state of ink absence in the pixel. If the pixel is not found to be within the blurred area, the ink presence/ink absence data is set for indicating the state of ink presence in the pixel.
If there exist plural blurred areas, that is if there exist plural sets of the blurred area data, the ink presence/ink absence setting unit 455 judges, based upon all of the blurred area data sets, whether or not the pixel detected at step ST4-44 is within the blurred area.
Then, at step S174-46, the first ink quantity setting unit 457 refers to writing brush data for setting ink quantity data in the pixel detected at step ST4-4-4 so that the shorter the length of the figure from the beginning point as far as the line segment li, the more becomes the ink quantity. However, if the ink presence/ink absence data specifying the state of ink absence in the pixel is set at step ST4-45, the first ink quantity setting unit 457 sets the ink quantity data to 0.
109 Then. at step ST4-4-7. the first ink density setting unit 456 refers to writing brush data for setting ink density data in the pixel det-ected at step ST4-4-4 so that the shorter the length of the figure from the beginning point as far as the line segment ii, the higher becomes the ink density. If, however. the ink presence/ink absence data is set at step ST4-4-5 for specifying the state of ink depletion in the pixel, the first ink density setting unit 456 sets the ink density data of the pixel to 0. After end of the step ST4-4-7, processing reverts to step ST4-4-2 to repeat the processing.
However, if there are plural pixels detected at step ST4-44, the ink presence/ink absence data, ink density data and the ink quantity data are set from pixel to pixel at steps ST4-45 to ST4-5-7.
On the other hand, the second pixel detection unit 453 detects, at step ST4-4-8, a pixel on a line segment 1 interconnecting the pixel on the line segment Ii for which the ink data has been set at steps ST4-4-5 to ST4-4-7 and the pixel on the line segment 1j., for which the ink data has been set at steps ST4-4-5 to ST4-4-7. The second picture data set-ting unit 454 calculates the ink density and the ink quantity in the detected pixels, based upon the ink data set on the pixel on the line segment li, ink data set on the pixel on the line seg-menz lj,,, distance between the detected pixel and the pixel on the line segment 1, and the distance between the detected pixel and 1 10 the pixel on the line segment li+1. for setting the ink density data and the ink quantity data on the detected pixels. After setting the ink density data and the ink quantity data in a-]] pixels on the line segment 1, the second pixel data setting unit 454 terminates the processing.
Since plural pixels on the line segment 1 are usually detected, the second pixel data setting unit 454 calculates the ink density and the ink quantity for each detected pixel for setting the ink density data and the ink quantity data.
If there are plural pixels on the line segment Ii or on the 1 ine segment l+li the second pixel detection unit 453 sets plural 1 i ne segments nte rconnect i ng the pi xe I s on t he 1 i ne segment and the pixels on the line segment 50 that these line segments extend along the contour of the f igure, f rom pixel to pixel, and detects pixels on the 1 ine segment I for each of these line segments -1. If the number of the pixels on the line segment li differs f rom that of the pixels on the 1 ine segment li.P the line segments are set so that the position of the pixels on the 1 ine segment li along the width of the f igure corresponds to that of the Pixels on the line segment li., along the width of the figure.
The above explanation has been made for the case of setting full-contor data representing an object of the blurred drawing, t h at is -ull-con-, our data selected as an object. of blurred drawing by the contour data selecting unit 422. If ink data is to be set for the fullcontour data not representing the object of blurred drawing, that is the full-contour data not selected by the blurred data selecting unit 422 as the object of blurted drawing. the ink data setting unit 45 disregards the ink presence/ink absence data, or assumes the ink Presence/ink absence data to be ink presence data, in setting the ink density data and the ink quantity data.
(4-4) Method for Setting Data Values Next, illustrative values of various data set in the blurred picture processor 40 are explained.
(4-4-1) Setting of Slurring Parameters by First Blurring Parameter Calculating Unit The first blurring parameter calculating unit 432 sets two sets of blurring parameters. The reason the two sets of the blurring parameters are set is that two blurred portions are produced at most per stroke when a character is actually written with the writing brush on the paper, so that it suffices to set two blurred areas for representing a blurred character. However, the number of sets of the blurred parameters is not limited to two and may naturally be varied depending upon the fig ure to be produced.
In the description to follow, the portion of the blurring data ti o.f the blurring writing brush carrying data in which-the value of ti is continuously " 1 " as shown at A in Fig.64 is termed.1 an area with t, = 1.0". Similarly, the portion of the blurring
1 1 2 data ti in which the value of ti is continuously ---0.5' as shown at S in Fig.64 is termed "an area with ti = 0.5", while the portion of the blurring data ti in which the value of ti -is continuously "0.3" as shown at C in Fig.64 is termed "an area with ti = 0.3".
The first blurring parameter calculating unit 432 operates in accordance with the flowchart shown in Figs.65 to 67 for setting the blurring parameters based upon the blurring brush carrying data.
At step ST4-5-1, the first blurring parameter calculating unit 432 initializes two sets of blurring parameters before transferring to step ST4-5-2. Specifically, the parameters-cl to c7 are al 1 set to 0.
At step ST4-5-2, the f i rst blurring parameter calculating unit 432 extracts the area with t, = 1.0. If there exists one or more of the area with ti = 1.0 (YES), the f i rst blurring parameter calculating unit 432 transfers to step ST4-5-3.if there exists no area parameter calculating At step ST4-5-3, unit 432 extracts the with ti = 1.0 (NO), the first blurring unit 432 transfers to step ST4-5-13.
k.
the firstblurring parameter calculating area with t, = 1.0. If there exist two or more of the area with ti 1.0 (YES), the first blurring paramete,r calculating unit 432 transfers to step ST4-5-4._ If only one such area exists, the first blurring parameter calculating unit 432 transfers to step ST4-5-6.
1 13 At step ST4-5-4, the f irst blurring parameter calculating unit 432 selects two areas with ti = 1.0 and transfers to step ST4-5-5. Two of these areas with ti = 1.0 are selected in the order of the decreasing length of the area with ti = 1.0, that is in the decreasing order of the number of ti in the area with ti At step ST4-5-5, the first blurring parameter calculating unit 432 sets two sets of blurring parameters in association with the two areas with ti = 1.0, selected at step ST4-5-4, in accordance with the equation (4-2), to terminate the processing.
cl = Tis-: Ti-l.; X Ti -1 C2. = Tie+: TI-1: XTI-1 2 C3 - C4 3 CS = e+ T;..: X (I-T.,) C6 = e±: Ti-i.: X c- = 1 -0. 08x (C2 -c.
l - .. (4-2) In the above equation, Ti stands for the value of tj in an area as an object of setting of the blurring parameters, that is 1.0, 0.5 and 0.3 for the areas with ti = 1. 0, t j = 0.5 and with ti = 0.3, respectively, and 1Til stands for the length of the area. The length of the area is normalized with the figure length being set to 1 and hence assumes the value of f rom 0 to 1. The. same holds for the length of the f igure, as explained subseque-tly.
On the other hand, Ti_1 specif ies the value of ti in an area 1 14 with ti =0.5 or 0.3 neighboring toone side of the area for which the blurring parameters are set. while 117j-11 denotes the length of the area. Similarly, Ti., specifies the value of ti-in an area with ti = 0.5 or 0. 3 neighboring to the other side of the area for which the blurring parameters are set, while 117j.11 denotes the length of the area.
Ti. stands for the length of the figure f rom the beginning point of the figure as far as the starting point of the areas with ti = 1.0, 0.5 or 0. 3 selected for setting the blurring parameters, while T i. stands for the length of the figure from the beginning point of the f igure as far as the end point of the areas with t,.0, 0.5 or 0.3 selected for setting the blu.rring parameters.
In the above equation, e is a coefficient representing an initial value of the fluctuations of the blurring start position and the blurring end position along the length of the f igure, and is pre-set to a value on the order of 1/20.
At. step ST45-6, the f i rst blurring parameter calculating unit 432 sets a set of blurring parameters, in association with the area with ti = 1.0, in accordance with the equati on (4-2). The f irst blurring parameter calculating unit 432 then transfers to step ST4-45-7.
At step ST4-5-7. the f irst blurring parameter calculating unit 432 extracts the areas with ti = C).5, If there is one or more area with ti = 0.5 (YES), the processing transfers to step 1 1 5 0 ST4-5-8.
ST4-5-1 0.
If otherwise (NO), the processing transfers to step At step- ST4-5-8, the f irst blurring parameter calculating unit 432 selects one area with ti _= 0.5 before transferring to step ST4-5-9. In selecting the area with ti = 1.0, Such area with the longest length of the area ti = 0.3, that is with the largest number of ti in the areas with ti = 1 0.
At step ST4-5-9, the f irst blurring parameter calculating unit 432 sets a set of blurring parameters in association with the areas with tj = 0.5 selected at step ST4-5-8, in accordance with the following equation (4-3):
cl= Tis-: Ti-l: XTi-1 C3 -C4 2 CS =e+: X C6 = a+ X C7 =0 -6 (43) before terminating the processing.
At step ST4-5-10, the first blurring parameter calculating unit 432 extracts the areas with ti = 0.3. It there is one or more area with t, = 0.3 (YES), the processing transfers to step ST4-5-11. If otherwise (NO), the processing is terminated At 5--ep ST4-5-11, the first blurring parameter calculating unit 432 selects an area with t, = 0.3 before transferring to 1 16 step ST4-5-12. In selecting the area with ti = 0.3,. such area with the longest length of the area ti = 0.3, that is with the largest number of ti in the areas with ti = 0.3, is selected.- At step ST45-12, the first blurring parameter calculating unit 432 sets a set of blurring parameters in association with the areas with ti = 0.5 selected at step ST4-5-11, in accordance with the following equation (4- 4):
(4-4) On the other hand, at step ST4-5-13 shown in Fig.66, the first blurring parameter calculating unit 432 extracts the areas with ti = 0.5. If there is one or more area with ti = 0. 5 (YES), the processing transfers to step ST4-5-14 and, if otherwise (NO), the processing transfers to step ST45-21.
At step ST4-5-14, the first blurring parameter calculating unit 432 extracts the areas with ti = 0. 5. If there are two or more areas with ti = 0.5 (YES), the processing transfers to step ST4-5-15 and. if otherwise (NO), the processing transfers to step ST4-5-17.
-ep ST4-5-15, the first blurring parameter calculating At s L unit 432 selects two or more areas with ti 0.5 before 1 17 transferring to step ST45-16. Two of these areas with ti = 0.5 are selected in the order of the decreasing length of the area with ti = 0.5-, that is in the decreasing order of the number -of ti in the area with ti = 0.5.
At step ST4-5-16, the first blurring parameter calculating unit 432 sets two sets of blurring parameters in association with two areas with ti = 0. 5 selected at step ST4-5-15, in accordance with the above equation (4-3), before terminating the processing.
At step ST4-5-17, the first blurring parameter calculating unit 432 sets a set of blurring parameters in association with the area ti = 0.5, in accordance with the above equation (4-3), before transferring to step ST45-18.
At step ST4-5-18, the first blurring parameter calculating unit 432 extracts areas with ti = 0.3. If there is one or more area withti = 0.5 (YES), the processing transfers to step ST4-519 and, if otherwise (NO), the processing is terminated.
At step ST45-19, the first blurring parameter calculating unit 432 selects an area with ti = 0.3 before transferring to step ST4-5-20. In selecting the areas with ti = 0.5, such area with the longest length of the area ti = 0.5, that is with the largest number of ti in the areas with ti = 0.5, is selected.
At step ST4-5-20, the first blurring parameter calculating unit 432 sets a set of blurring parameters in association with L he area wi t h t 0. 3 sel ec ted at step ST4-5- 19, i n acco rdance with the above equation (4- 4), before terminating the processing. 1 is At step ST4-5-21 shown in Fig.67, the f i rst blurring parameter
calculating unit 432 extracts the areas with ti = 0.3. If there is one or more area with ti = 0.3 (YES), the processing transfers to step ST4-5-22 and, if otherwise (NO), the first blurring parameter calculating unit 432 terminates the processing.
At step ST4-5-22, the first blurring parameter calculating unit 432 extracts two areas with t, = 0.3. If there are two or more areas with ti = 0.3 (YES), the processing transfers to step ST4-5-23 and, if otherwise (NO), the processing transfers to step ST4-5-25.
At step ST4-5-23, the first. blurring parameter calculating unit 432 selects two areas with ti = 0.3 before transferring to step ST4-5-24. Two of these areas with ti = 0.3 are selected in the order of the decreasing length of the area with ti = 0.3, that is in the decreasing order oil the number of ti in the area with ti = 1.0.
At step ST4-5-24, the first blurring parameter calculating unit 432 sets two sets of blurring parameters in association with two areas with ti = 0. 3 selected at step ST4-5-23, in accordance with the above equation (4-4), before terminating the processing.
At step ST4-5-25, the first blurring parameter calculating unit 432 sets a set of blurring parameters in association with the area ti = 0.3, in accordance with the above equation (4-4), to terminate the processing.
119 (4-4-2) Setting Blurring Parameters by Second Blurring Parameter Calculating Unit An illustrative embodiment of the blurring parameters set by the second blurring parameter calculating unit 433 is explained.
In setting the blurring parameters by the second blurring parameter calculating unit 433. the figures are classified, based upon the shape data, as shown for example in Tables 1 and 2:
TABLE 1
11 1 1 11 !i numb 1 1 1 1 eng th e r of of f igu bend re S 1 0 :. 1 111 1 1 i i 1 i f 1 4 i 1 1 J I i 4 i i 1 bend width of of f igure i 1 i 1 1 i i 1 1 - 1 i 1 1 length mean betwee widt n figure bend points 1 1 i i 1 1 1 h betw een bend poin ts patt ern 1 D3<r avg 1 Dasw min D <W max W mi n<DS D5 <W i max ern 2 D4:5W max 1 W-Max 1 pattern 3 W Max ' U patt-: 1 ern 4 i i r_avg 1 D <W :5D3 max 1 patt- - ern 5:
L:5 D D <L D:5W 4 max W max:5D5 W max<D, 1 1 1 1 i patt- ' ern 6 1 pattern 7 pattern 8 Li<2/3 D4:5W2 patt xL ern W2 <D 4 pattern 10 2/3xL patt :5LI ern num lengt length mean i h bend wi dt between width ber of of h bend f igur f igu of points between of ben e re figu bend re points 1 I 1 i i patt I i - i e r n 1 1/5xL<L W SW pat, t 2 W 2:5W 2 3 ern 1 2 W <wg p a 3 e r n 13 121 0 LSDI 4 r mor L i i i i 1 i 1 L2< 1 /5 XL L35L1 L2:51 /5 XL Li<L3 1/35L! 2-L3 i i L 4 < L 1 L I 5 L4 1/5xL<Li patt ern 14 W2 'S 1 W2:5W3 i W2 5W4 1 1 patt i ern 1 16 patt ern pat ern W4 <W 3 1 patt i m 1 r ern 16 patt:1 e r n 1 7 patt ' ern 18 patt ern 19 patt I 1 Li5 1 /5 xL patt 1 - 1 1 ern 123 In Tables 1 and 2, D1, 021 D31 D41 D5 and D, stand for values pre-set for classifying the figures.
In Tables 1 and 2.. the number of bend points is based upon bend paint data and stands for the number of portions bent in rnore than a pre-set angle. The length of the figure is based upon length data and L stands f or the length of the' f igure. The bend of the f igure is based upon curvature data, with r-avg representing the mean value of the curvature data. The figure width is based upon width data and W min stands for the smallest value of the figure width represented by the width data, while Wmax stands for the largest value of the figure width represented by the width data.
The length between bend points is based upon curved point data and the length data and stands for the figure length which is the figure length between bend points. It is noted thal, L stands for the length of the figure from the beginning point to the end point and Li stands for the f igure length f rom the (i1)'th bend point to the i'th bend point. That is, L, stands for the figure length from the beginning point to the first bend point and L2 stands for the figure length from the beginning point to the end point if the number of bend points is 1 and the figure length from the beginning point to the second bend point if the number of bend points is 2. On the other hand, L, Stands for the figure length from the second bend point to the end point if the number of bend points is 2 and the figure length f rom the 124 second bend point to the third bend point if the number of bend points is 3. while L4 stands for the figure length from the third bend point to the end point if the number of bend points is 3 and the figure length from the third bend point to the fourth bend point if the number of bend points is 4.
The mean width between bend points is a value based upon the curved point data and the width data and stands for the mean figure width between the bend points. It is noted that W, stands for the mean width of the figure from the i-l)'st bend point to the i'th bend point. That is, W, stands for the mean width of the figure from the beginning point to the end point, while Wz stands for the mean width of the figure from the bend point to the end point and that from the first bend point to the second bend point if the number of bends is 1 and not less than 2, respectively. on the other hand, W, stands for the mean width of the figure f ram the second bend point '&.a the end point and that f ram the second bend point to the third bend point if the number of bends is 2 and 3, respectively, whi le W4 stands for the mean width of the figure from the third bend point to the end point.
The second blurring parameter calculating unit 433 sets blurring parameters in accordance with the patterns classified as shown in Table 1 and 2 so that the blurring has character is1 ics as shown in Tables 3 and 4. Figs.68 to 87 illustrate examples of blurring when the blurring parameters are set so that the blurring will have characteristics shown in Tables 3 an 4.
position of blurred area patte mid part j r n patte rear par rn 2 L length of blurre d area about 1/3 about 1 1 1 1/3 ' i width manner of of blurred blurring area about 1/3 about il intense 1/2 exam ples 1 intense 1 Fig. 68 1 f i j F. ig. 69 1/2 to p a r t 1 y Fig.
patte end part 1/8 to 1/3 1.0 linear 70 r n and 3 partly intense 1/2 to near patte 1 rear part 1/4 to Fig r n 1/2 3/4 71 4 patte near bend j about 1/3 to i linear Fig.
rn point 1/3 1/2 72 end about about linear Fig.
patte near b rn 3/8 1/3 73 6 point patte end part about 1/3 to linear Fig.
rn 1/3 1/2 74 7 patte end part about about partly Fig.
rn 1/2 linear 75 8 1/4 and p a r t 1 y intense 126 patte near bendli abouti r n point 1/2XW2 i e I 1 1 end part about 1 2/3xW2 j, 1 / Z3 to 2/3 p a r t 1 y 1 i near and p a r t 1 y intense !1 F i 9 76 about intense Fig.
1/2 i 77 a Position length width manner of Ex Of of of blurring m blurred b 1 u r r e blurred pie area d area area patte end part about about partly -Fi 9 rn 2/3xW2 2/3 intense, 1 - and partly 78 linear J patte mid part about 1/3 to pa rt 1 y Fig linear rn 2/3xW 1/2 12 and partly 79 intense patte end part rn 13 - patte. fore half rn 14 end part about 1/2xW 3 about 1/2 to 2/3xW3 2/3 about about 1/2xW 1/2 i 1/3 to i i 1/2 patte f i nest about 112 to linear rn Portion 1 /4xWI 2/3 16 - 1 i near 1 i near 1 i F i 9 i F i 9 81 pa r T 1 y n e a r i and partly: 82 intense 1 F i 9 83 127 patte end part about 1/2 to 1 i near r n 3/4xW 2/3 Fig 4 17 84 patte fore half about about intense Fig 1/2xW 1/2 rn 18 85 patte end part,l about about linear Fig r n j 1/2xW4 1/2 19 86 tte end part about 1/2 to linear Fig 1 pa 1/4 2/3 rn 87 In Tables 3 and 4, the position of the blurred area indicates the Position of the blurred area along the figure length. The "fore half" means that the blurring occurs in the start point side region and the "rear half" means that the blurring occurs in the end point side region. The start point side area means a start point side area obtained on bisecting a figure into two equal portions along its length, while an end point side area means an end point side area obtained on bisecting a figure into two equal portions along its length. The start part, mid part and the end part mean the cases blurring occurs in the start area, mid area and in the respectively. The starting area means a start point obtained on tri-secting a figure into three portions 1 eng-L h. The mid area and the end area mean a mid point and an end point side area obtained on tri-secting into three Portions along its length, respectively.
1 2 8 in which end area, side area along its side area the f igure 0 The length of the blurred area means a mean length of the blurred area with respect to the figure length set to 1, while the width of the blurred area means a width of the blurred area with respect to the f igure width set to 1. W11 W2 and W3 mean the values of mean width between the bend points and are the same as those shown in Tables 1 and 2.
The manner of blurring means the blurring density. Thus, intense" means severe blurring in which only little ink is present in the blurred portion, while "linear- means weak blurring in which a considerable amount of ink is present in the blurred Portion and "partly intense and partly linear" means the moderate blurring intermediate between the intense blurring and linear blurring in which a certain amount of ink is present in the blurred portion.
(44-3) Setting of Ink Density Data by First Ink Density Setting Unit An illustrative example of ink density data as set by the first ink density setting unit 456 is now explained.
When setting ink density data id p on a pixel P on an i'th line segment li, the f i rst ink density setting unit 456 calculates and sets ink density data Idp, based upon ink densiiy 10 p of writing brush data associated with the drawing trajecto,ly passing tho-ught the pixel P, in accordance with the following ecuation (4-5):
Idp = I0P = a x i 129 ( 45) 0 where A is a coefficient of a pre-set value for representing the manner in which the ink density becomes gradually lower beginning f rom the start point Specif ical ly, it is set to a value on the order of 0.2.
However, if the value of Id p calculated by the equation (4is lower than a pre-set value, the f irst ink density setting unit 456 sets the ink density data Idp to 0. If the ink density becomes lower in this manner than he pre-set value, the figure is no longer drawn so that the blurring drawn is that near the end part of the figure.
On the other hand, the first in density setting unit 456 gradually sequentially modifies the ink density of the writing brush data for taking into account the ink flow within the writing brush during figure drawing. Specifically, it the values of ink density of the writing brush data corresponding to the k'th, (k-1)st and the (k+l)st drawing trajectories are I0k, iok-1 and IDk,l, respectively, the first ink density setting unit 456 before setting the ink density data Idp on apixel P onthe line segment Ii updates the ink density values of the writing brush data in accordance with the equation (4-6):
ID ID + b X (I0 IDO + b x(IDk k ki-1 - where is a coefficient of preset value for representing the ink f low within the writing brush. If IDk_1 is not present in the equation (4-6), that is if the ink density is the first ink density, I0k-; is set to ID If IDK.1 is not present k (I ok-1 o'.k) in the equation (4-6), that is if the ink density is the last ink density, IDk+1 is set to IDk (IDkl = IDk) When setting the ink amount data Iqp an the pixel P on the i 'th line segment ii, the f i rst ink quantity setting unit 457 calculates and sets the ink quantity data IqV based upon the ink quantity data IQ P of the writing brush data corresponding to the drawing trajectory passing through the pixel P, in accordance with the equation (4-7):
id q = IQ P x exp(c x i)...(4-7) where -Q is a coeff icient of a pre-set value for representing the manner in which the ink quantity becomes gradually thinner from the beginning Point. Specifically, it is set to a value on the order of -0.03.
However, if the value of Id P calculated by the equation (47) is lower than a pre-set value, the first ink density setting unit 457 sets the ink density data Idp to 0. Thus, if the ink density becomes lower than the pre-set value, the f igure is no longer drawn so that the blurring drawn is that near the end part of the figure.
On the other hand, the first in quantity s e t. t i n 9 -u n i t# 4 -5 7 gradual ly sequential 1 y modi flies the ink quantity of the writ ing brush data for taking into account the ink f low within the writing brush during f igure drawing. Specifically, if the values of ink quantity of the writing brush data corresponding to the k'l:h, (k-1)st and the (k+ l)st drawing trajectories are ICk, IQIC:
131 n, 1/ and IQk+1, respectively, the first ink quantity setting unit 457 before setting the ink quantity data Iq p an the pixel P on the line segment Ii updates the ink quantity values of the writing brush data in accordance with the equation (4-8):
I()k = Iak + d x (I0k+1 - I00 + d x (I0k-1 - I00 ..(4-8) where t is a coefficient of a pre-set value for representing the ink flow within the writing brush. If IQk-i is not present in the equation (4-8), that is if the ink quantity is the first ink quantity, ICk-1 is set to iak (1Qk-1 = Ok). On the other hand, if ICkf is not present in the equation (4-8), that is if the ink quantity'is the last ink quantity. I0k.1 is set to I0k (I0k+1 IQk).
(4-4-5) Setting of Ink Density by Second Ink Density Setting Unit An illustrative example of ink density data set by the second ink density setting unit 459 is now exPlained.
When setting ink density data Id, based upon the ink density data Id p of the pixel P on the line segment li, ink density data Id 1 of the Pixel Q an the line segment li, the distance PR between pixels Q and R on the line segment li and the distance OR between Pixels P and Q on the 1 ine segment I i+.,, the second ink density setting unit 458 calculates and sets the ink quantity data Iqr in =accordance with, for example, the equation (4-9) Idr Id p x (PR/P0) + Id X (OR/PG) ..(4-9) 132 where PQ stands for the distance between the pixels P and Q.
(4-4-6) Setting of Ink Quantity by Second Ink Quantity Setting Unit An illustrative example of ink quantity data set by the second ink quantity setting unit 459 is now explained.
When setting ink quantity data Iq, based upon the ink quantity data Iq P of the pixel P on the 1 ine segment 1,, ink quantity data Iq q of the pixel Q on the line segment 1 the distance PR between pixels Q and R on the line segment 1, and the distance QR between pixels P and () on the line segment li+,, the second ink quantity setting unit 459 calculates and sets the ink quantity data 1q, in accordance with, for example, the equation (4-9):
Iqr = lq P- x (PR/P0) + Iq x (QR/P0) ..(4-10) where PQ stands for the distance between the pixels P and Q. (4- 4-7) Setting of Luminance Data by Luminance Data Setting Unit An illustrative example of ink density data set by the luminance daza setting unit 46 is now explained.
* The luminance data setting unit 46 sets luminance- I in the pixel P, based upon paper f iber structure data formed by the paper f iber structure data forming processor 30 and picture data set by the first and second picture data setting units 454. That is, the luminance data setting unit 46 calculates and sets luminance 1 ineach pixel in accordance with the equations (4-11) 133 and (4-12):
I = 1 - f x Id p X Iq p (for M>6)... (4-11) I = 1 (for M:5 6)... (4-12) where Idp, Iq p and M stand for the ink density data, ink quantity data and the number of paper fiber structure data in the pixel p, respectively. The value of the luminance ranges from 0 and 1, with the initial value of the luminance I being 1.
In the above equations (4-11) and (4-12), _f is a coefficient pre-set to a Suitable value. If, in the equation (4-12), the value of I exceeds 1, the value of 1 is set to 1.
By setting the luminance I in this manner, it becomes possible to take account of the effect of paper fiber structure on blurring and hence to represent a blurred figure more realistically.
In the foregoing description, both the ink density data and the ink quantity data have been set as ink data. However, for simplifying data, only the ink density data or the in quantity data may be set. In such case, the luminance I in the pixel P may be calculated by simply modifying the equation (4-11) to the following equations (4-13) or (4-14):
I = 1 - f X Iq p (for M>6)... (4-13) 1 = 1 f x Id p (for M>6)...(4-14) (5) Bled Figure Drawing Processor Referring to Figs.88 to lil, a preferred embodiment of the bled figure drawing processor is explained.
134 If a f igure is drawn on a paper sheet with ink, bleeding is produced around a f igure. Based upon the modelled bleeding phenomenon, the bled figure drawing processor 50 forms picture data for representing a bled figure from picture data representing a f igure which is to undergo bleeding. The picture data of the bled f igure thus prepared by the blurred f igure processor 50 is fed to the display unit 1 for drawing a bled figure on the display screen of the display device 1.
For modelling the bleeding phenomenon, the information such as ink quantity or density in the figure to undergo bleeding is required. Thus, in preparing picture data of the bled figure in the bled figure drawing processor 50, the picture data of the figure to undergo bleeding includes data as to the ink quantity and ink density in the f i gure. Such picture data may be exemplif ied by the picture data formed in the blurred f igure drawing Processor 40.
(5-1) Modelling the Bleeding Phenomenon The bleeding generated an drawing a &figure on a paper sheet with ink is diffused by the ink f]owing in the interstices of the fibers making up the paper sheet. The interstices of the fibers through which flows the ink is referred to hereinafter as the capillary tube. The density of the color in the bled portion of the figure depends upon the ink absorbed by the paper sheet, that is by the quantity or the density o-1 the ink helc in the interstices oil the fibers making up the paper sheet. Thus, for 0 modelling the bleeding phenomenon, it is necessary to model the manner in which the ink f lows and to calculate the quantity and the density of the ink held in the fiber interstices. The modelled ink flow, calculation of the ink quantity held in the fiber interstices and calculation of the density of ink held in the fiber interstices will now be explained.
(5-1-1) Modelling the Manner in which Ink Flows The modelling of the ink flow is first explained.
Bleeding is diffused by the ink flowing in the capillary tube, as described above. Such ink flow may be modelled by assuming that the paper sheet is made up of plural micro-sized areas, referred to herein as pixels, the ink flows between the neighboring pixels if these pixels are interconnected by the capillary tube, and that the ink does not f low between the neighboring pixels if these pixels are not interconnected by the capillary tube.
The modelling may be performed by assuming that, when the ink f lows into a given pixel, the ink is held in the interstices of the f ibers in the pixel and absorbed therein, and that the ink is absorbed such that the ink that has not been absorbed by the pixel flows onto the next pixel interconnected with the firstly stated pixel by the capillary tube.
Thus t-he bleeding may be represented more realistically by n,odeling the ink flow by assuming that the ink flows through -,.he capillary tube interconnecting the pixels and is absorbed in the 136 pixel to which f lows the ink, with the ink absorbed in the different pixels producing the bleeding.
(5-1-2) Ink Quantity Held in the Fiber Interstices Referring to the drawings, the quantity of the ink held in the f iber interstices is explained. In the following description, the ink quantity held in the f iber interstices of a given pixel is calculated based upon the assumption that the above-described ink f low model is followed and the paper sheet is made up of plural fibers.
It is assumed that the fibers making up the paper sheet is of a twodimensional or planar structure and that the quantity of the ink held in the f iber interstices in a given pixel is proportional to the area occupied by the ink held in the fiber interstices of the pixel.
It is also assumed that there exist two f ibers FI, and FI, intersecting each other at an angle 6, as shown in Fig.88. The surface measure of the area occupied by the ink held in the interstices between the f ibers F1 and F2 is considered. The ink held in the interstices between the fibers FI, and FI2 is captured in an area of intersection between the two f.iberS FI1 and FI2 as indicated by hatching in Fig.88. The portion of t. h. e ink not contacted with the f ibers is bent under the surface tension of - the ink.
Thus the area of the ink held in the interstices of the two fibers FI, and FIZ may be approximated by an area surrounde_d by 137 the two f ibers FI1 and FI, and two circles each with a radius R, inscribing the two f ibers FI, and FI2. The radius R, is a constant which depends upon the surface tension of the ink.
If such approximation is used, the smaller the angle of intersection e between the two f ibers FI1 and FI2, the larger becomes the area occupied by the ink kept in the interstices between the two f ibers FI, and FI2. However, in eff ect, if the angle of intersection e between the two f ibers FI, and FI2 is smal ler than a pre-set value, the area occupied by the ink held in the interstices of the f ibers FI, and FI2 becomes smal]er.
Thus an approximation is made that, if the angle of intersection e between the two f ibers FI, and FI, is smaller than a pre-set angle e., the area occupied by the ink held in the interstices between the two f ibers FI, and FI, becomes larger. If the area occupied by the ink captured in the interstices between the two f ibers FI, and FI2 is approximated as described above, the of an area s(e) occupied by the ink cap-Lured in between t he two f i be rs FI, and FI2 i s 9 i ven as a angle e by the equations (5-1) and (5-2):
R2 x ccos ( e/2)/sin(e/2) + (e - TO/2) (for 6k ea)... ( 5 1) 2 x R2 X Ccos(e,/2)/sin(e,/2) + (e. - TO/2)/e (for e < e) surface measure the interstices 4 unction of thE s(8) = 2 x s(e) = If there exist n f iberS F11, FI2...... FIrl-1 1 138 1.. ( 5-2) FI, in a pixel p, as shown in Fig.90. The surface measure of the area occupied by the ink captured in the interstices between the n fibers FI 1 1 FIZ. FIn-1, FIn in the pixel P is found in the following manner. It is assumed that the angles of the n f ibers FI,, FI 2t FIn-1, FI areei, where i is 1, 2, 3,..., n-1, withthe f iber FI1 as a reference, and the angles of intersection of the neighboring f ibers are Aei, where i is 1, 2, 3,..., nl, as shown for example in Fig.90. Then, using the function s(e) as found by the equations (5-1) and (5-2), the area q(n) occupied by the ink captured in the interstices between the n fibers FIl. F12..... 3 FIn-11 FI. in the pixel P is found as the function of s(e) by the following equation (5-3):
lz it q(n) =1r-11X 6,) delae2---dez - 1 fl 0 0 0 .. (5-3) Thus the quantity U of the ink abSorbed in the pixel having n fibers is given by the equation (5-4) U= axq(r,) U= . (5-4) 139 0 where CL is a proportionality constant of the quantity of ink absorption U and the surface measure q(n) of the area occupied by the ink.
That is, if the quantity of ink absorption U in the pixel is found. it is possible to find the number n of the fibers in the pixel and the angles of intersection Aei between the neighboring fibers.
(5-1-3) Thickness of Ink Captured in Fiber Interstices Referring to the drawings, the concentration (thickness) of the ink captured in the fiber interstices is explained. In the following explanation, the shape of a capillary tube MO is assumed to be a cylinder through the inside of wh,ich the ink flows. The ink flowing direction through the cylinder of the capillary tube MO is indicated as x, the radial direction in the capillary tube MO with the center of the capillary tube MO as an origin is L and the time is t. The radius of the capillary tube mo is r,.
The ink is a dispersion of ink particles in a solvent, such as water, with the ink thickness being set by the number of ink particles. If the ink flows through the inside of the capillary tube, the ink particles become depleted towards the distal end of the flow. That is, when the ink f lows through the inside of the capillary tube, the ink f low become thinner towards its distal. For calculating the ink thickness which underc-ces f luctuations within the capillary tube, it is suf f icient if 0 d i f f u 5 i o n of the ink particles considered.
If the number Of diffused ink particles at time t and a position (x, r) is represented by n(x, r, t), the velocity Of the ink flowing in the direction x is v(r) and a constant representing the degree of diffusion in the direction _r of the ink particles is DO, the diffusion of the ink particles in the capillary tube may be represented by the following partial differential equation:
in the capillary tube is an (X, r, c) = D. x '-In (.r) X Vn lat ot dependent In the above equation, the. ink velocity v(r) is n upon the Position in the direction x but is dependent only upon the position in the r direction, such that the ink velocity is faster and slower towards the center and in wall surface of the capillary tube, velocity at a Position of r = 0 i maybe represented by, the vicinity of the respectively. It the S vmax, the ink velocity v(r) for example, the following equation (5-6):
v(r) = vmax x Cl (rlro)z) .. (5-6) The Partial differential equation (55) may be converted into the following difference equation (5-7), with a m inor variant of t, x and r being At, Ax and Ar:
141_ 0 n(x, r, t + At) n(x, r, t) At/Ax x v(r) x n(x-Ax, r, t) At/Ax x v(r) x n(x-Ax, r. t) 0 0 X At/Ar2 x n(x, r+Ar, t) + 00 X At/Ar2 x n(x, r-Ar, t) 2 x DO x At/Arz x n(x, r, t)... (5-7) In the equation (5-7), the left side denotes the number of ink particles at time t + At and at a position (x. r). The first term, second term, third term, fourth term, fifth term and the sixth term of the right side denote the number of ink particles at time and at a position (x, r), the number of ink particles at time and at the position (x, r) which flow out at time t + At in the direbtion 2L from the position (x, r). the number of ink particles at time t- and at the position (x-Ax, r) which flow out at time t+At towards the position (x, r), the number of ink particles at time t and at the position (x, r+Ar) which are diffused at time t+At towards the position (x, r), and the number of ink particles at time t and at the position (x, r) which are diffused at time t±At from the position (x, r) in the direction r, respectively.
If initial conditions are set for the equation (5-7), the numbers of particles n(x, r, t) at time _t and at a position (x, r) may be found sequentially. In sequentially calculating the number of ink particles n(x, r, t), it may be assumed that n(x, 1 A 2 ro, 0) = 0 since the ink particles reaching the wall surface of the capillary tube may be presumed to become affixed to the wall surface of the capillary tube without becoming diffused any further.
The initial conditions for solving the equation (57) may be set so that, when the ink starts to flow out, that is at t = 0, the ink particles are present only at the position x = 0 and are uniformly diffused in the r direction. That is, n(x, r, 0) =0 forx >Oand n(O, r, 0) =n,where n, is a constant.
It is noted that the sum of the ink particles in the r direction at the time and the position x may be represented by the sum of the number of ink particles R(x, t) flowing at time t and position -x and the number of the ink particles which have become affixed to the wall surface of the capillary tube at time t and position x, as shown by the equation (5-8):
V(x, t) = R(x, y) + Q(x, t) Thus the number of the ink particles R(x,flowing at time t and posi t i on _x may be calcul ated based upon n(x, r, t as shown by the equation (5-9):
Z, p X,.7) 2:r r x,-: (x, r, C) dr 0 n tb e other hand, the number of ink particles 0 (x. t) which have become affixed to the wall surface of the capilla,-y tube at time t and position 1 may be calculated based upon n(x, 143 t) and is given by the equation (5-1o):
c (x, y) = -2 ir ro xfDO n (x, z, z) & 0 .. (5-10) By the above calculations, it is possible to f ind the sum v(x, t) in the r direction of the numbers of ink particles at time 1. and position 2i. The ink thickness at time _t and position x is proportionate to V(x, t). Thus the ink thickness at time t and position x may be represented by multiplying V(x, t) with a suitable coefficient.
(5-2) stricture of Bed Figure Drawing Processor An illustrative structure of the bled figure drawing processor 50 -for forming picture data of the bled figure is explained.
Referring to Fig.92, the bled figure drawing processor 50 includes a contour pixel detection unit 51 for detecting the pixel P on the contour of a f igure yet to be bled, and an initial ink quantity setting unit 52 for pre-setting an initial ink quantity UG on the pixel P detected by the contour pixel detection unit 51. The bled figure drawing processor 50 also includes an initial in density setting unit 53 for pre-setting an initial ink thickness V, as picture data on the pixel P on the figure contour, and a 1 i nki ng pi xel detection unit 54 for detecting the pixels linked to the pixel P by the capillary tube based upon the number of capillary tubes Si (i = 1, 2, 3, 4, 5, 144 6, 7, 8) of the Paper fiber structure data previously formed by the paper f iber structure data forming processor 30 for the pixel p detected by the contour pixel detection unit 51. The bled f igure drawing processor 50 also includes a picture data setting unit 55 for setting data specif ying the ink presence for the pixel detected by the linking pixel detection unit 54 as picture data and a controller 56 for controlling the repetition of the operation of the coupling pixel detection unit 54 and the picture data setting unit 55. The f igure yet to be bled is e. g., a f igure represented by picture data formed by the blurred f igure drawing processor 40. That is, the bled f igure drawing processor 50 reads out picture data formed by the blurred figure drawing processor 40 and temporarily stored in the memory 3 for performing bled figure drawing processing on the read-out picture data.
Specif ical 1 y, picture data of a f igure to be bled is read from e.g., the input device 1, memory 3 and the external storage device 4 so as to be routed to the on-condor pixel detection unit 51 of the bled figure drawing processor 50.
The contour pixel detection unit 51 detects the pixel P on the contour of the figure based upon picture data of the figure yet to be bled. If there are plural pixels on the contour of the figure, the-contour pixel detection unit 51 sequentially detects these pixels. Meanwhile, such picture data of the fligure to be bled suffices which permits detection of the contour pixels. That 1 4 5 js, the picture data may be such picture data consisting only of the contour information, such as picture data specifying the Contour of a stroke of a character or a symbol based upon a font formed by the font forming processor 20.
The initial ink quantity setting unit 52 sets the initial ink quantity U, as picture data for the pixel P detected by the contour Pixel detection unit 51. Similarly, the initial ink density setting unit 53 sets the initial ink density VG as picture data for the pixel P detected by the contour pixel detection unit 51. The values of the initial ink quantity U 0 or the initial ink density V, may be pre-set values, prestored in the external storage device 4 or input values from the -input device 1 It is also possible for the information of the initial ink quantity UO or the initial ink density V, to be included in advance in picture data of a figure yet to be bled, while it is similarly possible for the initial ink quantity U, and the initial ink density V 0 to be set by the initial ink quantity U, or the initial ink density setting unit 53 based upon this information. Alternatively, the initial ink quantity Uo or the initial ink density VO may be pre-set in picture data of the figure yet to be bled without employing the initial ink quantity setting unit 52 or the initial ink density setting unit 52.
The pixel P, which has been detected in this manner by the contour Pixel detection unit 51 and which has the initial ink quantity Vo and the initial ink quantity V 0 thus set, is supplied 146 from the contour Pixel detection unit 51 to the linking pixel detection unit 54.
The linking pixel detection unit 54 detects pixels connected by the capillary tube to the pixel P detected by the contour pixel detection unit 51 based upon the number of capillary tubes Si ( i = 1, 2, 3, 4, 5, 6, 7, 8) associated with the pixel. Specifically. the number Si of the capillary tubes is data included in the paper fiber structure data formed by the paper fiber structure data forming unit 30 and specifically is an integer not less than 0.
If the number Si of the capi 1 lary tubes of the pixel P is not less than 1, the linking pixel detection unit 54 detect's the pixels Pi corresponding to the number Si of the capillary tubes as the pixels 1 inked to the pixel P. Specifically, if the numbers of the capillary tubes S1, S31 55 of the pixel P associated with directions of the pixels P,, P3 and PS neighboring to the pixel P are 1 and the numbers of the capillary tubes S2, S41 S6, S7 and S, of the pixel P associated with di rections of the pixels P21 P41 P61 P7 and P, neighboring to the pixel P are 0, the linking pixel detection unit 54 detects the pixels P1, -P3 and PS as pixels connected by the capillary tubes, as shown in Fig.93.
Alternatively, if the numbers of the capillary tubes Si of the pixel P-ares not less than 1 and the numbers of the capillary tubes S, 1, 2, 3, 4, 5, 6, 7, 8) of the pixels P, corresponding to the numbers S, are not less than 1, the linking 147 pixel detection unit 54 detects the pixel Pi as the pixel 1 inked to the pixel P. Specifically, if the numbers of the capillary tubes SI, S3. S5 of the pixel P associated with directions of the pixels P1, P3 and PS neighboring to the pixel P are 1, the numbers of the capillary tubes S2. S31 S60 S7 and S3 of the pixel P associated with directions of the pixels P2, P4, P61 P7 and P, neighboring to the pixel P are 0, the number S5 of the capillary tubes of the pixel P, corresponding to the direction of the pixel P is 1, the number S7 of the capillary tubes of the pixel P, corresponding to the direction of the pixel P is 1, and the number S1 of the capillary tubes of the pixel P5 corresponding to the direction of the pixel P is 0, the linking pixe,l detection unit 54 detects the pixels P1 and P, as pixels connected to the pixel P by the capillary tub.es, as shown in Fig.94.
The pixels connected to the pixel P by the capillary tubes, thus detected by the linking pixel detection unit 54, are routed from the linking pixel detection unit 54 to the picture data setting unit 55, The picture data setting unit 55 sets data indicating the presence of the ink for the pixels linked to the pixel.P by the capillary tube. If the initial ink quantity U, set on the pixel P is not more than 0, the picture data setting unit 55 assumes that there Is no ink diffused from the pixel P and terminates the bled figure drawing processing for the pixel P without setting data indicating the presence of the ink for the pixel connected 148 to the pixel P by the capillary tube.
In addition, the picture data setting unit 55 calculates, for each pixel for which the data indicating the presence of the ink is set, the quantity U of ink absorbed by the pixel, and sets the quantity U as picture data. If the quantity Q of the absorbed ink exceeds 0, the ink may be assumed to be present, the quantity Q of ink absorption being then simultaneously used f or setting the ink presence specifying data. The quantity U of the absorbed ink may be found, as explained in connection with the modelling of the bleeding phenomenon, in accordance with the equation (5-1l):
U= a XIT-1xl,f... fl: S (A ej d61C1192 0 0 0 2=1 TC Tr 11 n-1 where, if 0 k e, cos 2 e S.n( 2 E) - = 2 X'p2X S- (A0) = 2 ' and, if E) < E), 149 0 cos e,,) 6 -X s(A0) =2x.0,2x - - 2 + sin 6,) 2 .. (5-11) In the equation (5-11), n stands for the number of fibers M of the pixel which may be obtained by referring to the paper fiber structure data. The angle Aei is the angle of intersection between two neighboring fibers, that is between the i'th fiber and the (i+l)st fiber neighboring thereto, and is set at random in the picture data setting unit 55 within a range of 0:5 Aei:S Trso that thevalue of (. Ael + Ae 2 +.... ±AO n- 1)will be not larger than Tr. The value of R, which is a constant derived f rom the surface te.nsion of the ink, may be pre-set or entered by the input device 1 so as to be freely set by the user. The value of a, an optionally set proportionality constant, becomes larger or smaller if the absorbed ink quantity is larger or smaller, respectively. On the other hand, e a is a constant employed in a method for calculating the quantity of ink captured in the fiber interstices and may. for example, be Tr/4.
The value of U(n) is previously calculated for _n being an integer not less than 0, in accordance with the equation (5-11), and is previously set as a one-dimensional array of -n. When calculati'ng the absorbed ink quantity U, the picture data setting unit 55 refers to this pre-set onedimensional array in order to f ind the value Of the absorbed ink quantity U corresponding to the number M of the f ibers of the pixel. This signif icantl y reduces the processing time since there is no longer the necessity of calculating the absorbed ink quantity U in accordance with the equation (5-1 1) each time the absorbed ink quantity U is set.
The absorbed ink quantity U, calculated in accordance with the equation (5-1 1), is the maximum ink quantity absorbed by the pixel. that is the maximum absorbed ink quantity Umax. If the quantity of the ink f lowing to the pixel is smal ler than this maximum absorbed ink quantity Umax, the picture data setting unit 55 naturally sets the quantity of the ink f lowing to the pixel that is the ink quantity smaller than the maximum absorbed ink quantity Umax,- in the picture data as the absorbed ink quantity U.
The picture data setting unit 55 also calculates. for each pixel for which the ink presence indicating data has been set, the ink density v in the pixel, and sets the ink density V as the picture data. The ink density V is found in accordance with the equation (5-12):
V = V X V(x ' t (5-12) by calculating the number of ink particles V(x, t) at the time t and a -, a Position x based upon t he simulated bleeding phenomenon described above, 151 In the above equation, VO is the initial ink density set f or the pixel on the contour of a f igure yet to be bled, and x stands for the length of a capillary tube through which the ink f lows before reaching the pixel from the figure contour, specifically, the number of a series of pixels interconnected by the capillary tubes and sequentially detected by the linking pixel detection unit 54. On the other hand, t stands for the time elapsed until the ink reaches the pixel from the figure contour, specifically, the number of times of processing operations performed by the linking pixel detection unit 54 and the picture data setting unit 55.
Also, in the above equation, V(x, t) is pre-set as a twodimensional array of x and t based upon calculation af the equations (58) to (5-10) for the case of x and 1. being integers not less than 0. When f inding the ink density V, the picture data setting unit 55 performs the calculations oil the equation (5-12) in order to find the value of the ink density V by having reference to the value of V(x, t) among the pre-set twodimensional array V(x, t) corresponding to the number of times of repetition of the operations oli the linking pixel detection unit 54 and the picture data setting unit 55 until finding the number of a series of pixels interconnected by the capi 1 lary tubes and sequentially detected by the linking pixel detection unit 54 and the ink density V of the pixels. This significanTly reduces the processing time since there is no longer the 152 necessity of calculating v(x, t) in accordance with the equations (5-8) to (5-10) each time the ink density V is set.
The picture data setting unit 55 multiplies the absorbed ink quantity U with the ink density V thus found to f ind the luminance I of the pixel which is also set as picture data.
The bled figure drawing processor 50 routes the pixel, for which picture data have been set in the picture data setting unit 55, to the linking pixel detection unit 54 as the pixel P detected by the contour pixel detection unit 51 and iterates the operations of the linking pixel detection unit 54 and the picture data setting unit 55 for sequentially setting picture data simulating the gradual spreading apart of ink during bleeding.
During iteration of the operations of the linking pixel detection unit 54 and the picture data setting unit 55, new picture data are not set for a pixel for which picture data have already been set based upon the assumption that no new ink flows to the pixel to which the ink f lew previously. That is, no picture data is set by the picture data setting unit 55 for the pixel within an area occupied by the figure yet to be bled or the pixel for which picture data have already been set in the picture data setting unit 55.
The iteration of the operations by the linking p i x e 1 detection u-nit 54 and the picture data setting unit 55 is controlled by the controller 56.
That is, the controller 56 controls the iteration of. the 153 0 operations of the 1 inking pixel detection unit 54 and the picture data setting unit 55 based upon the assumption that bleeding comes to a close when the quantity of the ink diffused by bleeding has reached the initial ink quantity Uo which is the quantity of the ink responsible for bleeding.
Specifically, the controller 56 manages control so that the absorbed ink quantities U, calculated by the picture data setting unit 55, are summed for all pixels detected by the linking pixel detection unit 54, and the iterative operations by the linking pixel detection unit 54 and the picture data setting unit 55 is terminated when the sum value reaches the initial ink quantity U0- Alternatively, the controller 56 manages control so that the absorbed ink cruantities U, calculated by the picture data setting unit 55, is sequentially subtracted from the initial ink quantity U, set by the initial ink setting unit 52 for each pixel detected by the linking pixel detection unit 54, in order to find sequentially the residual i-nk quantity UCCSL this new ink residual quantity U,,,Z. is set as the new initial ink quantity UO in the pixel P detected by the contour pixel setting unit 51, and the iterative operation of the linking pixel detection unit 54 and the picture data setting unit. 55 is terminated when the new initial ink quantity U becomes not larger than 0.
1 0 If "the initial ink quant i t y ui differs between tw o neighbori ng pixels detected by the contour pixel detect ion unit 154 51, the controller 56 controls the iterative operation of the linking Pixel detection unit 54 and the picture data setting unit 55 after decreasing, for example, the initial ink quantity U, of the pixel with the larger initial ink quantity UO and increasing the initial ink quantity u, of the pixel with the smaller initial ink quantity UO for taking into account the ink flow on the contour of the figure yet to be bled. Since the ink flows only gradually with lapse of time, the operation of decreasing or increasing the initial ink quantity U 0 by the controller 56 as described above is to be carried out gradually at the time of the iterative operation by linking pixel detection unit 54 and the picture data setting unit 55.
In the above-described embodiment, the setting of the ink density V and'the luminance I by the picture data setting unit is carried out on the occasion of the iterative operations by the linking pixel detection unit 54 and the picture data setting unit 55. If, however, only the bled figure is displayed on the display device 5, it suffices if the ink density V and the luminance I on termination of ink diffusion by bleeding is known, so that it suffices to set the ink density V and the luminance I at the time of the last iterative operation by the 1 inking pixel detection unit 54 and the picture data setting unit 55.
If onl.
y the bled figure is displayed on the display device 5, the routing of picture data f rom the picture data setting unit 55 to the display device 5 is performed on termination of the iteration of the operation of the linking pixel detection unit 54 and the Picture data setting unit 55, that is after the end of formation Of Picture data by the bled figure drawing processor 50. However, if the manner of spreading apart of ink during bleeding from the figure yet to be bled is sequentially displayed on the display device 5, the picture data setting unit 55 sends the picture data to the display device 5 each time the operation of the linking pixel detection unit 54 and the picture data setting unit 55 is iterated, that is each -time picture data is set in the picture data setting unit 5. That is, the picture data setting unit 55 routes the data of luminance I to the display device 5 for each iteration of the operations of the linking pixel detection unit 54 and the picture data setting unit 55 for sequentially d-isplaying the figure derived from the luminance I on the display device 5. Thus the manner of gradual spreading apart of ink during bleeding from the figure yet to be bled may be sequentially displayed on the display device 5.
(5-3) Bled Figure Drawing processing - An illustrative operation of the bled figure drawing processor 50 is now explained. In the following description, the terms used in bled figure drawing and data used in the bled figure drawing processing are first explained and the opera-Lion of the bled figure drawing processor 50 is then explained by referring to a flowchart illustrating the operation. (5-3-1) Terms Used in Bled Figure Drawing processing
156 In the following description, the pixels corresponding to the contour of a figure are termed "contour pixels". Bleeding is produced by diffusion of the ink from the contour pixels via the pixels interconnected by the capillary tubes. Among the contour pixels, the contour pixel from which originates the bleeding to a pixel A, lying on the downstream side along the path of ink diffusion, that is the contour pixel reached when the ink flow is followed retrogressively from the pixel A,, is termed "-Lhe contour Pixel from which originates the bleeding towards the pixel A 11. ' or "bleeding originating pixel with respect to the pixel P".
If, during bled figure drawing processing, two or more downstream side pixels interconnected by the capillary tubes with the pixel AI are found, the p ixel A, is termed a "branching point pixel". Specifically, if two pixels BP c,, on the downstream side of the pixel A, are detected as the pixels interconnected by the capillary tubes, the pixel A, is termed the branching point pixel.
If the ink flows to the pixel AP the branching point pixel which is reached when the ink flow path is followed retrogressively from the pixel A, and from which the ink flows in a direction other than the direction towards the pixel AP is termed "the-. ink flow branching point pixel with respect to the A series of pixels interconnected by the capillary pixe I A," - L.
tutes and lying on an ink path from the pixel AI to the ink 157 branching point Pixel with respect to the pixel At, are termed "a series of pixels f rom the pixel AI t the ink f low branching point pixel ". If, when the i nk f low path i s ret rogressivel y f ol lowed f rom the pixel A,. there exists no i nk f low branching point pixel with respect to the pixel A,, a series of pixels f rom the pixel AI to the bleeding originating contour pixel, interconnected by the capillary tube, are termed "a series of pixels f rom the pixel AI to the ink flow branching point pixel".
Thus, if there exist a series of pixels A,, A2 and A3 interconnected by a capillary tube, and a series of pixels C,, C2 and C3 interconnected by the capillary tube, the pixel A3 is connected to the pixel C, by a capillary tube and the ink f lows within the series of the pixels Cl to C5 interconnected by the capillary tube, as shown in Fig.96, the pixel C3 is the ink f low branching Point Pixel with respect to the ink f low branching paint pixel, while the pixels AI 1 A2 and A, are the series of pixels f rom the pixel AI to the ink f low branching point pixel.
If there exist a series of pixels All A2 and A3 interconnected by a capi 1 lary tube, a series of pixels C1 j C21 C31 C4 and CS interconnected by a capillary tube, and a series of pixels DI, D2, D3i D4 and D5 interconnected by a capillary tube, the pixel A, is connected to the pixel C3 by a capillary tube, the pixel C5 is connected to the pixel D, by a capillary tube, the ink flow in the pixels C, and C, ceases and the ink i'lows within the series oil the pixels 0, 1.0 05 interconnected by the 158 capillary tube, as shown in Fig.97, the pixel 03 is the ink f lew branching point Pixel with respect to the piXel A;, whi le the pixels AI ' A21 A3, C31 C4 and CS are the series of pixels f rom the pixel AI to the ink f low branching point pixel.
Also. if there exist a series of pixels AI ' A2 and A, interconnected by a capi 1 lary tube, a series of pixels Cl 1 C2, C3, C4 and CS interconnected by the capillary tube and a bleeding originating contour pixel D, the pixel A3 is connected to the pixel C3 by a capillary tube and the pixel CS is connected to the bleeding originating contour pixel D by a capillary tube, with the ink not flowing through the pixels Cl and C2, as shown in Fig.98, the pixels A,, A21 A3, C3, C4 and CS are the seri. es of pixels f rom the pixel A, to the ink f low bra nching point pixel. (5-3-2) Data Employed in Bled Figure Drawing Processing In the following descr iption, an area for drawing a bled f igure is a planar area consisting of plural pixels P, as shown in Fig.99. The paper fiber structure data employed has at least a number M of f ibers and a number of capillary tubes S1 j S21 S3j S41 S51 S j S, and S in association with respective pixels, as shown in Fig. 100.
The bleeding processing by the bled f igure drawing processor 50 employs a number of times of processing flag for recording the nurnbe r of. times of processing, contour data set f or correspondence to pixels on a con- Lour of a figure (contour pixels) yet to be bled, current data set in association with 159 respective Pixels positioned at the forefront of bleeding diffusion at a given time point, next data set in association with respective pixels positioned a t the f o ref ront of the bleeding diffusion at a time point next to the time point when the current data is written, bled picture data set in association with respective pixels constituting an area for bled f igure drawing and luminance data set in association with respective pixels constituting an area for drawing a bled figure.
The contour data is made up of contour pixel coordinate values, an initial ink quantity U, of each contour pixel and an initial ink density V, for the contour ink, as shown for example in Fig.101. Each contour data has a contour data number.
The current data has a coordinate value of a pixel A at the forefront portion of bleeding diffusion at a given time paint, referred to hereinafter as the current leading coordinate value, a coordinate value of a pixel neighboring to the pixel A and operating as an ink originating pixel for the pixel A, referred to hereinafter as the current bleeding originating coordinate value, a contour data number for the contour data for the bleeding originating contour pixel for the pixel A, and the status specifying the ink status in the pixel A, as shown for example in Fig.102. Each current data has the current da:a number. Since the current data is the data for a pixel positioned at the foremost portion oil bleeding diffusion a! a given time point, the current data is sequentially rewritten wiih progress in the bleeding diffusion, The next data has a coordinate value of a pixel S Positioned at the foremost Portion of the bleeding diffusion at. a time point next to a time Point when data is written in the current data, referred to hereinafter as the next leading coordinate value, a coordinate value of a pixel neighboring to the pixel B and operating as an ink originating pixel for the pixel 8, referred to hereinafter as the next bleeding originating coordinate value, a contour data number for the contour data for the bleeding originating contour pixel for the pixel B, and the status specifying the ink status in the pixel 9, as shown for example in Fig.103. Each next data has the next data number. Since the next data is the data for a pixel positioned at the foremost portion of bleeding diffusion at a time point next to the time point when data is written at the current data, the current data is sequentially rewritten with progress in the bleeding diffusion.
Similarly to the paper fiber structure data, the bleeding picture data is associated with each pixel constituting an area for writing a bled figure. The bleeding picture data for a given pixel P has a status specifying the ink status at the pixel P, the number of steps specifying the number of times of processing required since the start oil bleeding diffusion until inflow oil the ink to the pixel P, -he number of steps for ink flow st-op specifying the number of times of processing required since the 161 start of bleeding diffusion until the stop of the ink flow to the pixel P, a coordinate value of a pixel neighboring to the pixel P and operating as an ink flow originating pixel for the pixel P, referred to hereinafter as a bleeding originating coordinate value, a contour data number of contour data associated with the bleeding originating contour pixel for the pixel P. an absorbed ink quantity U specifying the quantity of the ink absorbed by the pixel P, and an ink density V specifying int density at the pixel P, as shown for example in Fig.104.
In the statuses of the current data, next data and the bled picture data, "no ink- specifying the ink absent state as an initial value, originating" specifying that the pixel P corresponds to a bleeding originating pixel, "flowing" specifying that the ink is flowing to and flowing through the pixel P, and "stop" specifying that the flow to the pixel has ceased, are set.
Similarly to the paper fiber structure data, luminance data corresponds to the pixels constituting an area for drawing a bled figure. The luminance picture data for the pixel P has luminance for the pixel P, as shown for example in Fig.105.
(5-3-3) Operation of Bled Figure Drawing processor An illustrative operation of the bed f igure drawing processor 50 is now explained by referring to the flowchart of Figs.106 to 110.
First, at step ST5-1, picture data of a figure to be bled, 162 for example, Picture data formed by the blurred figure drawing processor' 40, is read from, for example, the memory 3 or the external storage device 4. The contour pixel detection unit 51 sets the status of bled picture data corresponding to all pixels on the f igure to be bled to "originating" based upon picture data of the figure to be bled. The processing then transfers to step ST5-2.
At step ST5-2, the contour pixel detection unit 51 judges whether or not there exists any contour pixel not processed by steps ST5-3 to 5-5. If there is any such pixel (YES) processing transfers to step ST5-3 and, if ot he rwise (NO), processing transfers to step STS-56.
At step ST5-3, the contour pixel detection unit 51 detects a contour pixelnot processed by steps STS-3 to 5-5, and forms contou r dat a f o r t he con tou r pixel. The contour pixel detection uni t 5 1 af f i xes a contou r dat a numbe r t o eac h cont ou r dat a, wh i 1 e forming contour data using the coordinate values of the contour pixels as the contour coordinate values.
At step ST5-4, the initial ink quantity setting unit 52 sets an initial ink quantity U, in the contour data formed at step ST5-3.
At step ST5-5, the initial ink density setting unit 53 sets an initia 1 ink density V, in the contour data formed at step ST53. After step ST5-5, the processing reverts to steo ST5-2 to repeat the processing.
163 0 By repetition of the steps ST5-2 to ST5-5, contour data corresponding to the contour pixels is formed with a unique contour data number being affixed to each contour data. These contour data may be pre-set in picture data of the f igure to be bled such that the bled figure processor 50 simply reads the pre-set contour data without performing the processing by the steps ST5-2 to ST5-5.
At step ST55-6,' the controller 56 sets the value of the number of times of processing flag f to 0.
At step ST5-7, the linking pixel detection unit 54 judges whether or not there is any contour data not processed by steps ST5-8 to ST5-10. If there is any such pixel (YES), the proce ssing transfers to step ST5-8 and, if otherwise (NO), the processing transfers to step ST5-11 shown in Fig.107.
At step STS-8. the linking pixel detection unit 54 detects a contour data not processed by steps ST5-9 to ST5-10. The contour data detected by this step ST5-8 is referred to hereinafter as "contour data to be processed", while the pixel co r respond i ng to i he contour data to be p rocessed i s ref er red to as "contour pixel to be processed---. The processing then transfers to step ST5-9.
At step ST5-9, the linking pixel detection unit 54 refers to the paper f iber structure data in order to detect a pixel interconnected to the contour pixel to be processed by a capillary tube. The pixels to be detectecl, are pixels neighboring 164 to the Contour Pixel being processed, with the number of the detected Pixels being 0 to 8. The pixels for which the bleeding data has the status "Originating", that is pixels on or inside of the contour of the f igure yet to be bled, are not detected. The number of pixels to be detected is eight when there is only one contour pixel. Usually, the number of pixels to be detected is 0 to 7.
At step ST5-10, the linking pixel detection unit 54 forms current data associated with the pixel detected at step ST59. Specifically, the linking pixel detection unit 54 forms the current data, with the current data number affixed thereto, so that the status is "flowing", the coordinate value of a pixel detected at step ST5-59 is a leading coordinate value, the coordinate value of a contour pixel of contour data 'La be processed is an ink flow originating coordinate value and the con- Lour data number affixed to the contour data to be processed is the contour data number. Since the number of the pixels detected at step STS- 9 is 0 to 8, the number of current data formed at step ST5-10 is not limited to 1 but ranges from 0 to 8. The number of pixels detected by step ST510 is likely to become eight when there is only one contour data. The number of the current data formed at step ST510 is usua 1 1 y 0 to 7. After step ST5-10, the processing reverts to step ST5-7 in order to repea! the processing.
BY repetition of the steps S75-7 to ST5-10, the initial 0 current data is formed with a unique current data number being affixed to each current data.
At step ST5-1i, shown in Fig.107, the controller 56 increments the value of the number of times of processing flag f by 1.
At step ST5-12, the picture data setting unit 55 judges whether or not there is any current data not processed by steps ST5-13 to ST5-27, If there is any such current data (YES), the to step STS-13 and, if otherwise (NO), the to step ST5-24 shown in Fig.108.
the picture data setting unit 55 detects a contour data not processed by steps ST5-14 to ST5-27. The current data detected by this step ST5-13 is referred to hereinafter as -current data to be processed", while the pixel corresponding to the current data to be processed is referred to as "current pixel to be processed". The processing then transfers to step ST5-14.
AT step STS-14, the picture data setting unit 55 calculates the maximum absorbed ink quantity Umax of the current pixel to be processed. The processing then transfers to step STS-15.
At step ST515, the picture data setting unit 55 judges. based upon the maximum absorbed ink quantity Umax, calculated at step STS-14,- the absorbed ink quantity U set in the bled picture data corresponding to the current pixel to be processed, and the initial ink quantity u, of the contour data corresponding to the processing transfers processing transfers At step ST5-13 166 contour data number of the current data to be processed, whether the ink flows in a quantity exceeding the ink quantity absorbable by the pixel or in a quantity not exceeding the ink quantity absorbable by the pixel. Thus the picture data setting unit 55 transfers to step ST5-16 and to step ST5-22 for (Umax - U) < U 0 and for (Umax U9): U 01 respectively. If the absorbed ink quantity U is not set in the bled picture data corresponding to the current pixels to be processed, the value of the absorbed ink quantity U is set to 0.
At step ST5-16, the picture data setting unit 55 refers to the status of the current data to be processed. If the status of the current data to be processed is "flowing", the processing transfers to step ST5-17 and, if the status of the current data to be processed is "stop", the processing transfers to step ST519.
At step ST517, the picture data setting unit 55 sets bled picture data corresponding to the current pixel to be processed. Specifically, the picture data setting unit 55 sets the bled picture data by setting the, status, the value of the f lag 1, the ink flow originating coordinate value of the current data to be processed, the contour data number of the current data to be processed, and the maximum absorbed ink quantity Umax calculated at step ST5-14 less the current absorbed ink quantity of the Lled picture data U (Umax - U) to "flowi,,)9", the number o,'i. s'Le,-s, the ink flow originating coordinate value, con-Lour data number 167 and to a new absorbed ink quantity U, respectively.
Then, at step STS-18, the controller 56 sets the current initial ink quantity U, of the contour data corresponding to the contour data number of the current data to be processed less a value obtained on subtracting the current absorbed ink quantity U of the bled picture data f rom the maximum absorbed ink quantity Umax calculated by step STS-14 (Umax - U), or (UO - Umax + U), as the new initial ink quantity UO of contour data corresponding to the contour data number of the current data to be processed. After step STS-18, processing reverts to step ST5-12 to repeat the processing.
At step ST5-19, the picture data setting unit 55 sets bled picture data corresponding to the current pixel to be processed. The picture data setting unit 55 sets the status, the value of the flag f., the ink flow originating coordinate value of the current data to be processed, the contour data number ofF, the current data to be processed, and sum of the initial ink quantity u, of the contour data corresponding to the Contour data number of the current data to be processed and the current absorbed ink quantity U of the bled picture data (U, + U), to "s-,-op", the number of steps, the ink flow originating coordinate value, the contour data number and a new absorbed ink quantity U, respectively.
At step ST5-20, as a-, step STS-18, the controller 56 se:s the current initial ink quantity UG of the con-Lour data 168 0 corresponding to the contour data number of the current data to be processed less a value obtained on subtracting the current absorbed ink quantity U of the bled picture data from the maximum absorbed ink quantity Umax calculated by step STS-14 (Umax U), that is (U0 Umax + U), as the new initial ink quantity U, of contour data corresponding to the contour data number of the current data to be processed.
At step ST5-21, the picture data setting unit 55 sets the statuses of all bled picture data corresponding to a set of pixels from the current pixel to be processed to the ink flow branching point pixel to "stop", while setting the flag f to the number of ink flow stop steps. After step ST5-21 the processing reverts to step ST5-12 to repeat the processin 9.
At step ST5-22, the pic ture data setting unit 55 sets bled picture data corresponding to the current pixel to be processed. The picture data setLing unit 55 sets the status, the value of the flag f., the ink flow originating coordinate value of the current data to be processed, the contour data number of the current data to be processed, and sum of the initial ink quantity UG of the contour data corresponding to the contour data number of the current data to be processed and the current absorbed ink quantity U of the bled picture data (UJ + U), to "flowing", the number of steps, the ink f low originating coordinate value, the contour data number and a new absorbed ink quant i t Y U, respectively.
value oil the 169 0 At step STS-23, the controller 56 sets the value of the initial ink quantity u, of the contour data corresponding to the contour data number of the current data to be processed to 0. After step ST5-23, the processing reverts to step ST5-12 to repeat the processing.
By repetition of the steps ST5-12 to ST5-23, bled picture data are set on the pixels to which the ink f]owed due to bleeding.
At step STS-24, shown in Fig.108, the controller 56 refers to the value of the initial ink quantity Uo of the contour data. If there is still the contour data exceeding the value of the initial ink quantity U, (NO), the processing transfers to step STS-25. If the value of the initial ink quantity Uo of all contou r data 15 not mo re t han 0 (YES), t he p rocess i ng t ransf e rs to step ST5-38 shown in Fig.110.
At st ep ST5-25, t he cont ro 1 1 e r 56 d imi n i s hes f 1 uctuat i ons in the value of the initial ink quantity UO of the contour data in order to take into account -he ink flow within the contour pi.xels. That is, if the value of the initial ink quantity UG of the contour data differs between two neighboring contour pixels, the controller 56 manages control for decreasing the value of the initial ink quantity U, of the contour data having a larger value of the initial ink quantity U, and for increasing the value of the initial ink quantitLy U, of -he contour data having a smialler value of the initial ink quantity U0. In decreasing the initial ink quantity U0, the minimum value of the ink quantity UO is set to 0 - At step ST5-26, the linking pixel detection unit 54 judges whether there is any current data not processed by steps ST5-27 to ST5-33. If there is such current data (YES), the processing transfers to step ST5-27. If there is no such current data (No), the processing transfers to step ST5- 34 shown in Fig.109.
At step ST5-27, the linking pixel detection unit 54 detects a current data not processed with steps ST-28 to ST5-33. The processing then transfers to step STS-28.
At step ST5-28, the linking pixel detection unit 54 refers to the bled Picture data set in the pixel corresponding to- the current data detected at step ST5-27, in order to compare the absorbed ink quantity U of the bled picture data to the maximum absorbed ink quantity Umax in the pixel. If the ink can be further absorbed, that is if U < Umax, the processing transfers to step ST5-29. If the ink cannot be further absorbed, that is if U = Umax, the processing transfers to step ST5-30.
At step ST529, the linking pixel detection unit 54 forms next data based upon the assumption that the pixel corresponding,o the current data detected at step ST5-27 is at the forefront portion of the bleeding diffusion. Specifically, the linking pixel detection unit 54 forms the next data, with the next data number aff ixed thereto, by setting the status, the current coordinate value of the current data detected at step ST5-27 and 171 the current ink f low originating coordinate value of the current data detected at step ST5-27 to "flowing", the next leading coordinate value of the next data, and to the next ink f low originating coordinate value for next data, respectively. Af ter step ST529, the processing reverts to step ST5-26 to repeat the processing.
At step ST5-30, the linking pixel detection unit 54 refers to the paper fiber structure data in order to detect the pixel linked via the capillary tube to the pixel corresponding to the current data detected at step ST5-27. The processing the transfers to step ST5-31. The pixels to be detected are seven pixels, SPeCifically, the pixels neighboring to the -pixel corresponding to the current data detected by the step STS-27, excepting the pixel from which the ink flow to the pixel under consideration originated. The number of the pixels detected at step STS30 is 0 to 7. However, the pixels whose bled picture data has the status "originating", that is pixels on and inside of the contour of the figure from which originates the bleeding, and Vh,e pixels 4Mcse bled picture data has the status "flowing or "stop", that is the pixels to which the ink flowed previously, are excluded from detection.
At step STS-31, the linking pixel detection unit 54 judges whether or not the pixels have been detected at step ST5-30. If the pixels have been detected (YES), the processing transfers to step ST5-32. If the pixels have not been detected (NO), the 172 U processing transfers to step ST5-33.
At step ST5-32, the linking pixel detection unit 54 forms next data, corresponding to the pixel detected at step ST5-30, based upon the assumption that the ink flow is continuing. Specifically, the linking pixel detection unit 54 forms the next data. with the next data number affixed thereto, by setting the status, the coordinate value oil the pixel detected at step ST5 30, the current leading coordinate value of the current data detected at step ST5-27 and the contour data number of the current data detected at step ST5-27, to "flowing", next leading coordinate value, next ink flow originating coordinate value and to the contour data number, respectively. Since 0 to 7 pixels are detected at step ST5-30, the number of the next data to be formed at step ST532 is not limited to 1, but may range from 0 to 7. After the step ST5-32, the processing reverts to step ST526 to repeat the processing.
At step STS-33, the linking pixel detection unilt 54 forms next data based on the assumption that the ink flow has ceased. Specifically. the linking pixel detection unit 54 forms the next data, with the next data number affixed thereto, by Setting the status, the coordinate value of the current data detected at step ST5-27, the current ink flow arig- inating coordinate value of Zhe current data detected at step ST1527 and the contour data number of t h e current data detected a: step STS-27, to "s',op", next leading coordinate value, next ink flow originating coordinate 173 0 value and to the contour data number, respectively. After the step ST5-33, the processing reverts to step ST5-26 to repeat the processing.
By the repetition of the steps ST5-26 to 5-33, the next data, which is data concerning the pixel positioned at the forefront of bleeding diffusion at a time point next following a time point when data is written in the cur-rent data, is formed, with the next data number proper to each next data affixed thereto.
At step STS-34, shown in Fig.109, the linking pixel detection unit 54 judges whether or not there is the next data. The presence or absence of the next data specifies the continuation and end of the bleeding diffusion, respectively. If there is the next data (YEES), the processing transfers to step ST5-35 and, if otherwise (NO), the processing transfers to step ST5-38.
At steps ST5-35 to ST5-37, the linking pixel detection unit 54 copies next data in the current data to update the forefront portion of the bleeding diffusion. That is, at step STS-35, the linking pixel detection unit 54 erases al current data. -Then, at step STS-36, the linking pixel detection unit 54 directly copies all current next data in the current data. At step STS-37, the linking pixel detection unit 54 erases all current next data. After step ST537, the linking pixel detection unit 54 - Lransfe.-s to step STS-11 to repeat the setting of the bled picture data.
174 0 On the other hand, at step ST5-38, the picture data setting unit 55 sets the bled picture data based on the assumption that the bleeding diffusion has ceased. That is, the picture data setting unit 55 sets the status of the bled picture data, whose status is Previously set to "flowing", to "stop", while setting the number of ink flow stop steps to the value of the flag f.
Then, at step ST539, the picture data setting unit 55 calculates the ink density V of a pixel to which f lowed the ink, that is a pixel whose bled picture data has the status "stop", based upon the number of ink flow stop steps and the number of steps of the bled picture data of the pixel, and sets the calculated-value of the ink density as the ink density V of the bled picture data. Specifically, the picture data setting unit 55 previously calculates the ink density V(x, t) for x and being an integer of not less than 0, for setting as a twodimensional array of x and t, in accordance with the equations (5-8) to (5-10). The picture data setting unit 55 ref ers to the ink density V(x, t) thus set as the two-dimensional array in order to find the ink density V. Tha-, is, the picture data setting unit 55 refers to the ink density V(x, t) set as a twodimensional array, with the number of steps and the number of steps of ink flow stop being 1 and t, respectively, for flinding the ink density at each pixel.
At step ST5-40, the picture data setting unit 55-5 calculazes the luminance value for each pixel, based upon the bled picture data set for each pixel, for setting luminance data. The luminance value of a pixel to which f]owed the ink, that is a pixel whose bled picture data has a status "stop", is calculated by multiplying the absorbed ink quantity U at the pixel with the ink density. For calculating the luminance value, the luminance value may be multiplied with a suitable coefficient for normalization. The luminance value of a pixel an a figure yet to be bled, that is a pixel whose bled picture data has the status 'originating", is set to a maximum value within the gamut of the luminance value, while the luminance value of a pixel to which no ink f lowed, that is a pixel whose bled picture data has the status "stop", is set to 'originating", is set to a maximum value within the gamut of the luminance value. It is assumed in the foregoing that the larger the luminance value, the greater become the ink quantity and ink density. However, the processing nay naturally be performed on the assumption that the smaller the luminance value, the greater become the ink quantity and ink density.
At step STS-41, the bled picture drawing unit 50 routes the luminance data set for each pixel at step ST5-40 to the display device 5. Based upon the luminance data, the display device 5 displays the bled figure an the display device 5.
In the above-described bled figure drawing processing, the bled picture data is set cin the assumption that no ink 1 lows fron, a pixel to a given pixel to which the ink previously flowed from 1 76 0 another pixel. That is, when detecting the pixel connected by a capillary tube at step ST5-30, any pixel to which the ink previously f]owed, that is a pixel the status of the bled picture data of which is "flowing" or "stopped", is excluded from detection.
In effect, however, ink particles are likely to be diffused to a pixel, to which the ink previously flowed from a given pixel, from another pixel interconnected to the pixel by a capillary tube.
Thus the bled picture data may be set on the assumption that, if one or more pixels connected by capillary tubes to the pixel P to which the ink previously flowed exist besides the pixel f rom which originated the ink f low to the pixel P, and the ink previously flowed to such one or more pixels interconnected by the capillary tubes to the pixel P, that is if the ink flow at the pixel P col 1 ides with the ink f low at the one or more pixels connected to the pixel P by the capillary tubes, ink particles become diffused between the pixel P and the one or more pixels, and hence need to ',-ie taken into account in setting the bled picture data.
The method for setting the bled picture data for taking into account the diffusion of the ink particles in case of such collision of the ink flows is now explained by referring to Fig.11.
in an embodiment shown in Fig.111, there exist a series of 177 pixels AI, A2, A31, A42, A52 interconnected by a capillary tube, a series Of pixels S 11 B2, B31 B42, B52 interconnected by a capillary tube, and a series of pixels AC, A51, A61, B51, B 41 interconnected by a capillary tube, the pixels A, is connected by a capillary tube to the pixel A41 and the pixel A, is connected to the pixel B41, the ink flows through the pixels AP A 21 A31, A4Z, A52, in this order, as a f i rst ink f low, this ink f low is branched at the pixel A3 to flow towards the pixels A 41, A51, AC the ink also f lows through the pixels B1 9 B21 B31, B42, BR, in this order, as a second ink flow and the ink flow is branched at the pixel B3 to f low towards the pixels BSI and B41 In such case, the ink f lowing to the pixel A61 col 1 ides against the ink f]owing to the pixel B,,. For setting the bled picture data for respective pixels in such case, it suffices to compare the pressure of the ink f low at a pixel at the branching point to the pixel A61, that is at the pixel A3, to the pressure of the ink flow at a pixel at the branching point to the pixel B,P that is at the pixel B3, and to assume that, if the two pres sures are equal to each other, there is produced no ink particle diffusion, whereas, if the two pressures are not equal to each other, diffusion between the two particles is continuing.
Since the pressures of the two ink f lows at two pixels are larger at the initial stage of bleeding diffusion and become smal]er as the bleeding dif fusion progresses, it suff ices to compare the number of steps as set in the bled picture data for the pixel A3 and the number of steps as set in the bled picture data for the pixel 83 for comparing the ink flow pressures at the two pixels.
If the number of steps as set in the bled picture data for the pixel A3 and the number of steps as set in the bled picture data for the pixel B3 are equal to each other, no diffusion of the ink particles occurs at the pixels AW A51, A619 851, B41' Thus, for these pixels AW ASI, A61, BSI, B, P bled picture data are set in accordance with the above-described flowchart.
Conversely, if the number of steps as set in the bled picture data for the pixel A3 is smaller than the number of steps as set in the bled picture data for the pixel B3, diffusion of the ink particles occurs at the pixels A4P A51, A61, BSI, B4, in- direction of from the pixel A,, to the pixel BM until the first ink f low ceases. Thus, for the number of ink flow stop steps for the bled picture data of these pixels A A AC, B,-,, B,,, the 41, 51, same value as that of the number of the ink f low stop steps for the pixel A3, which is the ink flow branching point pixel for these pixels, is set.
On the other hand, if the number of steps as set in the bled picture data for the pixel B3 is smaller than the number of steps as set in the bled picture data for the pixel AS, diffusion of the ink particles occurs at the PiXelS A41' ASI, A61, BSI, B,, in a direction of from the pixel B,, to the pixel A,., un-,il the second ink f low ceases. Thus, for the number of ink flow stop steps for 179 the bled picture data of these pixelS A41, A51. A61, B51, BW the same value as that of the number of the ink flow stop steps for the pixel A3, which is the ink flow branching point pixel for these pixels, is set.
Thus the bled picture data may be set with the diffusion of ink particles in case of collision of different ink streams taken int consideration.
Although the example of two ink streams from two directions has been described in the foregoing, the number of ink flow stop steps for bled picture data in case oil collision of three ink streams from three different directions may similarly be set by comparing the number of steps of the bled picture data as set in the ink flow branching point pixels with respect to the pixel at which the ink flow collision has occurred.
(6) Figure Display Example Fig.112 shows an example in which a Japanese kanji character reading "yume" (meaning "dream"), made up of ten strokes YUl to Yulo, as an example of a blurred and bled figure, is actually displayed in the above- described picture display system.
In the example shown in Fig.112, the paper fiber st.-ucture data forming processor 30 forms paper fiber structure data, while the kanii character yume " is f o r m e _d b y the font f c rt-.i i ng processor 30. The font reading "yume" is made up of ten stroke data corresponding to the ten stroke data YUl to YU10.
The paper fiber structure data flernied by the paper fiter structure data forming processor 30 and the font reading 1. yume formed by the font forming processor 20 are read into the blurred figure drawing processor 40 which then performs blurred figure drawing processing on the respective strokes VUl to YU10 on the kanji character based upon these paper f iber structure data and the font of the kanji character in order to set the ink quantity and ink density in the strokes YUl to YU10 at a time point of drawing of the respective strokes YUl to YWO.
The ink density and the ink quantity 'of the writing brush data, which are data representing the state of the writing brush at the time point of writing the respective strokes YUl to YWO, are assumed to be the same at all points on the writing brush. The ink quantity and the ink density of the writing brush data, associated with the strokes YUl to YWO, are set as shown in Table 5.
TABLE 5 writing brush data ink quantity ink density Yul 0.50 0.45 YU2 0.43 0.45 strokes YU3 0. 10 0.45 YU4 0. is 0.45 YU5 0.47 0. 45 YU6 0.47 0.45 181 1 i YU7 0. is 1 0.45 1 H YU8 Yug 0.15 c).45 0.10 0.45 Yula 0.30 0.45 The blurring parameters for the strokes YUl to YU10 were set as shown in Table 6.
t 5, TABLE 6 r 0 i 1 k 1 e: S1 y 1 no U! blu 1 1 i 1 y:
I; U! A 1; 2! 1 1 1, 1 i if i :1 0 1 i:
i 1 C 2 rre d are a 1 0 1 i 0 1 i 1 i 1 1.0 1 1.0 0 0 1 1 i C6 c 7 C3 C4 C5 1.0: 0 1 1 c fir 0.5 0.7 0.0 0.0 1/1 6 0 0 28 7 0 st 1 blu 1 i 1 1 1 i 1 rre d i are 1 1 1 a 1 1 182 i 1 0 1 0 1 0 0 0 1 1 1.0 0 i 1 i i 1 0 sec 4/5 1 o 0.0 1 /2 0.0 1 /1 0.2 0 o 0 9 and blu r r e d are S y f i r 1/3 0.6 0.0 1.0 0.1 1/6 1/3 U 3 0 0 3 st 3 blu rre i d are a q l! ii !1 a sec 2/3 and blu rre d are a y f i r U st 4 blu r r e d are a 1 0.7, 0.0 1 6 0 1 1 i 1 1 0.5 0.7 0.0 6 0 0 28 1 1/2 0.1 0. 1 3 3 i i i 0.4! 1 i 1 0.0 1 / 1 1.0 7 0 0 i sec 4/5 1.0 1.0 1/2 0.0 1 /1 0.2 0 0 0 9 and blu r r e d a r e 1 1JL3 0 1.0 0 0 1.0 1.0 i 1.0 0 0 1.0 0 1 0 1.0 0 0. 1 1/3 0.0 1 0 7 0 0 1 1 i 1 1 1 1 1 1.0 1. 0 1 o 0 0 0 0.1 3 i - 0.2 0.8 0.0 1/3 1/6 3 3 0 1 no 1 0 1 0 U blu 0 0 rre d are a no U bl U 6 rre d are a y f i r U st 7 blu r r e d are a y f r U st blu r r e d are a sec 0.4 i 3 and, blu, r r e U are a 1.0 0 0.8 0.1 6 1 i 1 0,7 0,0 1/2 1 /1 0. 1 6 0 0 3 184 0.4 2 0.3 ii 'I 1 0 Y! U Y no U blu 1 r r e 01 d l are a i f i r 0.4! 1 4 1 1 st blu r r e d are a sec 9/1 0 ond blu r r e d are a 1 0 0.6 1 0.0 0 i 0 5/6 0.0 0.1 7 0 0 1 i 1 i 1 1 C v 1 I 1.0 0.0 0 0 1 1 i 1 i 0.1! 3 8/1 0.0 5 1 0 1.0 0 i 1 i i 1.0 1.0 1.0 1.0 0 0 0 0 1.0 1. c) 0 i ? 1 1 0 1 The paper f iber structure data, formed by the paper f iber structure data formingprocessor 30, and the ink quantity and the ink density in the strokes YUl to YWO, as set in the blurred figure drawing processor 40, were read into the bled figure processor 50. The bled figure processor 50 performs bled figure processing on the respective strokes YUl to YU10 of the kanji character for producing bleeding around the strokes YUl to YU10 for drawing the kanji character reading "yume" which is blurred, and bled, as shown in Fig.112.
Thus it becomes possible with the picture display systern 0 to express blurring and bleeding which appear as if the character were actually written with a writing brush.
Meanwhile, it may occur that, when the figure is enlarged and displayed on the display device, the figure contour becomes jagged, as shown in Fig. 113. Thus, in computer graphics, antialiasing is carried out for smoothing the figure contour for displaying the smoothed figure on the display device.
The above-described bled figure drawing processing is also effective as such antialiasing processing. That is, by processing the vicinity of the contour potion with bled figure drawing processing, the figure contour may be smoothed, as shown in Fig. 114. In addition, with the antial iasing processing, employing the bled f igure drawing processing, the f igure contour may be displayed on the display device so that the figure will appear more like a figure actually drawn with a writing brush than a figure drawn with antialiasing capable only of smoothing the f igure contour. The reason is that the contour of the f igure actually drawn with the writing brush is not perfectly smooth but undergoes some bleeding.
186 0 1. A method for representing a blurred f igure on a display device based upon picture data setin a plurality of pixels in a figure, comprising:
reading first contour data including one or more coordinate po i nt s Aj ( i = 1, 2, 3,...) 1 y i ng on a f i r st cont ou r ex t end i ng from a beginning point to an end point of the figure and second contour data including one or more coordinate points Bj (i = 1, 2, 3,...) lying on a second contour extending from said 1 187 beginning point to said end point; detecting a pixel P lying on a line segment 1 k interconnecting a k'th coordinate point Ak of said first contour data and a k'th coordinate point B, of said second contour data; judging, based upon blurred area data specifying a blurred area of the f igure, whether or not said pixel P is within a blurred area; setting, if said pixel P is within the blurred area, ink absence data specifying the ink depleted state as Picture data in the pixel P, and setting, if said pixel P is not within the blurred area, ink presence data specifying the ink carrying state is set as picture data in the pixel P; detecting a pixel a on a line segment Iki-1 interconnecting (k+l)st coordinate point Ak1 of said first contour data and a (k+l)st coordinate point Bki.1 of said second contour data; judging, based upon sad blurred area data, whether or not said pixel Q is within the blurred area; setting, if said pixel 0 is within the blurred area, ink absence data specifying the ink depleted state as picture data in the pixel Q, and setting, if said pixel 0 is not within the blurred area, ink presence data specifying the ink carrying state as picture data in the pixel Q; detecting a pixel R lying an a line segment interconnecting said pix'ets P and Q; and setting picture data in said pixel R based upon respec"r,iive 188 picture data set in said pixels P and Q.
2. The blurred f igure drawing method as claimed in claim 1 further comprising supplementing a coordinate point between two neighboring coordinate points Ai and Aj., of said f i rst contour data, if the distance between said coordinate points Ai and Ai.1 is larger than a preset value, so that the distance between the neighboring coordinate points will be not larger than the pre-set value, for forming new first coordinate data, or supplementing a coordinate point between two neighboring coordinate points Band Bj.1 of said second contour data, if the distance between coordinate Points Bj and Bj., is larger than a preset value, so that the distance between the neighboring coordinate points will be not larger than the pre-set value, for forming new second coordinate data; and detecting said pixels P and 0 using said new first and second contour data.
3. The blurred f igure drawing method as claimed in claim further comprising supplementing coordinate points to said -first contour data or to said second contour data, if the number of the coordinate points of said f irst contour data differs f rom the number of the coordinate points oil said second contour data, so that the number of the coordinate poi nts of said f i rst contour data becomes equa I to the number of the coordinate points c', said second contour data, for forming new first contour data or new second contour data; and detecting said pixels P and Q using said new first and second coordinate data.
189 0 4. The blurred f igure drawing method as claimed in claim 1 wherein said Picture data has ink density data specifying the ink density. said ink density data being set in said pixel P so that the ink becomes thicker the shorter a figure length from said beginning point to the line segment lk, said ink density data being set in said pixel Q so that the ink becomes thicker the shorter a figure length from said beginning point to the line segment lk+l- 5. The blurred figure drawing method as claimed in claim 4 further comprising setting ink density data in the pixel R based upon ink density data set in the pixel P, ink density data set in the pixel Q, distance between the pixels P and R and the distance between the pixels R and Q.
6. The blurred figure drawing method as claimed in claim 1 wherein said picture data has ink quantity data specifying the ink quantity, said ink quantity data being set in said pixel P so that the ink quantity becomes larger the shorter the length of a figure from said beginning point to the line segment lk, said ink quantity data being set in said pixel Q so that the ink quantity becomes larger the shorter the length of a figure from said beginning point to the line segment 1k4-1, 7. The blurred figure drawing method as claimed in claim 6 further comprising setting ink quantity data in the pixel R based upon ink quantity data set in the pixel P, ink quantity data set in the pixel 0, distance between the pixels P and R and the distance between the Pixels R and Q.
8. The blurred figure drawing method as claimed in claim 1 further comprising:
referring to pixel-based paper fiber structure data including the number of fibers passing through each pixel; calculating luminance of a pixel for which picture data has been set based upon said picture data and the number of fibers for the pixel for which the picture data has been set; and representing a blurred f igure on a display device based upon calculated luminance.
The blurred f igure drawing method as claimed in claim the 9. wherein said blurred area data is comprised of m blurring start lengths representing the lengths on M trajectories extending through the f igure f ram a starting point. to an end point thereof substantially parallel to said first or second contour which are started f ram said starting point and terminated at blurring start points where a blurring area starts, and m blurring terminating lengths representing the lengths on said m trajectories starting f ram said starting point and terminating at blurring end points where the blurring area is terminated.
10. The blurred f igure drawing method as claimed in claim 9 further Comprising judging the pixel P ta be within and outside the blurred area if PS < L < PE and L, 5 PS or PE 5 L k k k, respectively, and judging the pixel 0 to be within and outside 79 7 the blurred area if QS < Lk+1 < OE and Lk+1 5 QS or QE:5 Lkl, respectively, where PS denotes the blurring start length of the blurring area data on a trajectory passing through the pixel P, PE denotes the blurring end length of the blurring area data on a trajectory passing through the pixel P, QS denotes the blurring start length of the blurring area data on a trajectory passing through the pixel Q, QE denotes the blurring end length of the blurring area data on a trajectory passing through the pixel Q, respectively, and Lk, LkH denote the f igure lengths f rom said starting point to line segments lk and 1k+1, respectively.
11 - The blurred f igure drawing method as claimed in claim wherein there are plural blurred areas in a f igure, a plurality of blurred area data are provided in association with each blurred area, and wherein it is judged whether or not. the pixels P and Q are within the blurred area f or each blurred area data.
12- The blurred f igure drawing method as claimed in claim 1 wherein said blurring parameters include a parameter c, representing a blurring start position along the length of said f igure f rom said starting point to said end point, a parameter C2 representing the blurring end position along the length of the figure, a parameter C3 representing the upper blurring end position along the width of the figure, a parameter c4 representing the lower blurring end position along the width of the f igure, a parameter c5 representing the extent. oil fluctuations of the blurring start position along the length 192 of the f i gure, a parameter cS representing the extent of fluctuations of the blurring end position along the length of the figure, and a parameter cT representing the blurring density along the width of the figure, said blurred area data being previously formed based upon said blurring parameters.
13. The blurred figure drawing method as claimed in claim 12 further comprising pre-setting a plurality of blurring parameters having different value combinations of said parameters c,, c2, c3, C4, cS, c6 and c,, and forming said blurred area data based upon blurring parameters selected from among said plural blurring parameters.
14. The blurred figure drawing method as claimed in claim 12 further comprising calculating said parameters CP C2, C3, C4, C5, c6 and C7 based upon speed distribution data representing writing brush carrying speed distribution data and writing brush pressure distribution data representing the writing brush pressure distribution during drawing the f igure f rom the start. point to the end point thereof, and using the calculated parameters c,, C2, C31 C41 C51 c, and C, as said blurring parameters. 15. The blurred f igure drawing met-hod as claimed in clairr. 12 further comprising calculating said parameters c,, C2, C3, C4, CS, c, and c7 based upon shape data specif ying the shape of a f igure, said shape data having length data from the beginning point of 193 said figure to an end Point thereof, width data specifying the width of the f igure, curvature data specifying the bend of the f i gure and curved poi nt data speci f yi ng the number and t he positions of bends at an angle larger than the pre-set angle; and using the calculated parameters c,, IC,, C31 C4' C5, c6 and C7 as said blurring parameters.
16. The blurred f igure drawing method wherein there are present a plurality of said f igures, as claimed in claim 1, f urther comprising:
selecting at least one of said figures in accordance with a pre-set rule; using the selected f igure as a f i gure yet to be blurred; and setting picture data for producing blurring for only pixels within the f igure yet to be -blurred while setting picture data for not producing blurring for pixels within the remaining figures not to be blurred.
17. The blurred f igure drawing method as claimed in claim 16 wherein said preset rule states that a _f igure be selected f rom among said plural f igures so that the ratio of the number of the selected blurred figures to the totality oil the figures is not larger than a pre-set value.
18. The blurred f igure drawing method as claimed in claim 16 wherein said pre-set rule states that a figure be selected from among said plural f igures so that the ratio of the number of the selected blurred figures to the totality of the figures is not 194 larger than a pre-set value, and that one of the remaining figures be selected at random.
19. The blurred figure drawing method as claimed in claim 17 wherein said Pre-set ratio is 1/3.
20. The blurred figure drawing method as claimed in claim 17 wherein the sequence of said plural figures is pre-set and wherein said Pre-set rule states that such a figure is selected from among said plural figures which has the largest figure length and longer figures are sequentially selected from among the remaining figures excluding both figures lying ahead and at back of said longest figure.
21. An apparatus for representing a blurred figure on a display device based upon picture data set in a plurality of pixels in a figure, comprising:
means for reading first contour data including one or more coordinate points Ai (i 1, 2, 3, lying on a first contour extending from a beginning point to an end point of the figure and second contour data including one or more coordinate points B (j = 1, 2, 3,...) lying on a second contour extending from said beginning point to said end point; first pixel detection means for detecting pixels on a line segment Ii interconnecting a coordinate point A, of the first contour d-ata and a coordinate point Bi of the second contour data based upon the first contour data and the second contour data supplied from said data readout means; first Picture data setting means for judging, based upon blurred area data specifying a blurred area of the f igure, whether or not the pixel detected by said f i rst pixel detection means is within a blurred area; setting. if said pixel is within the blurred area, ink absence data specifying the ink depleted state as Picture data in the pixel, and setting, if said pixel is not within the blurred area, ink presence data specifying the ink carrying state as picture data in the pixel; second pixel detection means for detecting a pixel positioned intermediate between the pixel lying on said line segment Ii set by said f i rst pixel setting means and a pixel lying on the line segment li,,; and second picture data setting means for setting picture data on the pixel detected by said second pixel detection means based upon the picture data of the pixel lying on the line segment li detected by said f i rst picture data setting means and the picture data of the pixel lying on the line segment li.I. 22. The blurred f igure drawing apparatus as claimed in claim 21 wherein said data readout means further comprises contour data supplementing means for supplementing a coordinate point between two neighboring coordinate points Aj and A,,, on said f i rst contour data so that the distance between the neighboring coordinate points will be not larcer than a pre-set value, if said distance is larger than the pre-set value, for producing new 196 first coordinate data, said contour data supplementing means for supplementing a coordinate point between two neighboring coordinate points Bi and Bi+l on said second contour data so that the distance between the neighboring coordinate points will be not larger than a pre- set value, if said distance is larger than the pre-set value, for producing new second coordinate data.
23. The blurred figure drawing apparatus as claimed in claim 21 wherein said data readout means includes contour data adjustment means for supplementing coordinate points to said first contour data or to said second contour data, if the number of the coordinate points of said first contour data differs from the number of the coordinate points of said second contour data, so that the number of the coordinate points of said first contour data becomes equal to the number of the coordinate points of said second contour data. for forming new first contour data or new second contour data.
24. The blurred figure drawing apparatus as claimed in claim 21 wherein said picture data has ink density data specifying the ink quantity, and wherein said f i rst picture data setting means has first ink density setting means for setting ink density data in the pixel detected by said f i rst pixel detection means so that the ink density becomes higher the shorter the length of the figure from said starting point to said line segment li.
25. The blurred fligure drawing apparatus as claimed in claim 21 wherein said second picture data setting means for ca iculating 197 the ink density in the pixel detected by said second pixel detection means based upon the ink density data set for pixels on two neighboring line segments Ii and 1j+1 by said first ink density setting means. distance between the pixel on the line segment Ii and the pixel detected by said second pixel detection means and the distance between the pixel detected by said second pixel detection means and the line segment li+l for setting ink density data on said pixels.
26. The blurred figure drawing apparatus as claimed in claim 21 wherein said picture data has ink density data specifying the ink quantity. and wherein said first picture data setting means has first ink density setting means for setting ink density data in the pixel detected by said f i rst pixel detection means so that the ink quantity becomes larger the shorter the length of the figure from said starting point to said line segment 11.
27. The blurred f igure drawing apparatus as claimed in claim 26 wherein said second picture data setting means for calculating the ink quantity in the pixel detected by said second pixel detection means based upon the ink quantity data set for pixel on two neighboring 1 ine segments Ii and li., by said f i rst ink quantity setting means, distance between the pixel on the 1 ine segment Ii and the pixel detected by said second pixel detection means and the distance between the pixel detected by said second pixel detection means and the line segiment flor setting ink quantity data on said pixels.
198 28. The blurred f igure drawing apparatus as claimed in claim 21 further comprising luminance data calculating means for referring to pixel-based paper fiber structure data including the number of fibers passing through each pixel for calculating luminance of a pixel for which picture data has been set based upon said picture data and the number of f ibers for the pixel for which the picture data has been set; a blurred f igure being represented on a display device based upon the pixel-based luminance data set by sad luminance data calculating means.
29. The blurred f igure drawing apparatus as claimed in claim 21 wherein said blurred area data is comprised of m blurring start lengths representing the lengths on m trajectories extending through the f igure f rom a starting point to an end point thereof substantially parallel to said f irst or second contour which are started f rom said starting point and terminated at blurring start points where the blurring area starts, and m blurring terminating lengths representing the lengths on said m trajectories starting from said starting point and terminating at blurring end points where the blurring area is terminated".- 3o. The blurred f igure drawing apparatus as claimed in claim 29 wherein the pixel P is judged to be within and outside the blurred area if PS < L k < PE and Lk:s PS or PE 5 Lk, respectively, and judging the pixel 0 to be within and outside the blurred area i -I- CS < Lkl < QE and Lk,-,:5 CS or GE 5 L-k-1, respecTively, PS denotes the blurring start length of the blurring area data on 199 a trajectory passing through the pixel P, PE denotes the blurring end length of the blurring area data on a trajectory passing through the pixel P. QS denotes the blurring start length of the blurring area data on a trajectory passing through the pixel Q, QE denotes the blurring end length of the blurring area data on a trajectory Passing through the pixel Q, respectively, and Lki LW denote the figure lengths from said starting point to line segments lk and lk+I, respectively.
31. The blurred figure drawing apparatus as claimed in claim 21 wherein there are plural blurred areas in a figure, a plurality of blurred area data are provided in association with each blurred area, and wherein said first picture data setting means judges whether or not the pixels detected by said f i rst pixel detection means for each blurred area are within the blurred a r e a.
32. The blurred figure drawing apparatus as claimed in claim29 further comprising blurring parameter setting means for setting blurring Parameters comprised of a parameter cl representing a blurring start position aleng the length of said figure from said starting Point to said end point, a parameter c2 representing the blurring end Position along the length of the f igure, a parameter c. representing the upper blurring end position along the width o, the f i.gure, a parameter c, representing the lower blurring end position along the width of the figure, a p a r a m e t e r CS representing the extent of f luctuations of the blurring s+ Lat position along the length of the f i gure a parameter c representing the extent of fluctuations of the blurring start position along the length of the f igure, and a parameter c, representing the blurring density along the width of the figure, and blurred area data forming means for forming said blurred area data based upon the blurring parameters set by said blurred parameter setting means.
33. The blurred f igure drawing apparatus as claimed in claim 32 wherein said blurring parameter setting means has blurring parameter selecting means for selecting one blurring parameter from among a plurality of preformed blurring paramet,:rs having different value combinations of said parameters Cl, C2, C3, C4, C5. c6 and C7.
34- The blurred f igure drawing apparatus as claimed in claim 32 wherein said blurring parameter setting means has f irst blurring parameter calculating means for calculating said parameters cl, C2, c3, C4, c5, c6 and c7 based upon speed distribution data representing writing brush carrying speed distribution and writing brush Pressure distribution data representing the writing brush Pressure distribution during drawing the figure from the start point to the end point thereof for setting said blurring parameters.
35. The. blurred f igure drawing apparatus as claimed in claim 34 wherein said input device is a tablet and wherein he move:-,,ent speed and the writing pressure of the pen are entered as said 201 speed distribution data and as said writing brush. pressure distribution data. respectively.
36. The blurred figure drawing apparatus as claimed in claim 32 wherein said blurring parameter setting means has second blurring parameter calculating means for calculating said parameters c 11 Cv cP C4, CS, c6 and c7 based upon shape data specif ying the shape of a f igure having length data f rom the beginning point of said figure to an end point thereof, width data specifying the width of the figure, curvature data specifying the bend of the f i gu re and cu rved po i nt dat a spec i f yi ng t he numbe r and the positions of bends at an angle larger than the pre-set angle f or setting the blurring parameters.
37. The blurred f igure drawing apparatus as claimed in claim 21 further comprising contour data readout means for reading a plurality of fullcontour data made up of first contour data and second contour data; and contour data selecting means supplied from said contour data readout means for selecting at least one flull-contour data from the supplied plural full-contour data in accordance with a pre-set rule; a figure represented by the full contour data selected by said contour data selecting means being blurred in figure drawing and a figure represented by the fullcontour data not selected by said contour data selecting means not being blurred in figure drawing.
38. The blurred figure drawing apparatus as claimed in claim37 wherein said pre-set rule states that a figure be selected from 202 among said plural figures so that the ratio of the number of the selected blurred figures to the totality of the figures is not larger than a preset value.
39. The blurred f igure drawing apparatus as claimed in claim 37 wherein said pre-set rule states that the full-contour data be selected from among said plural full-contour data so that the ratio of the number of the selected full-contour to the totality of the full-contour data is not larger than a pre-set value, and that one of the remaining full-contour data be selected at random.
40. The blurred figure drawing apparatus as claimed in claim 37 wherein said pre-set ratio is 1/3.
41. The blurred f igure drawing apparatus as claimed in claim 37 wherein the sequence of said plural full-contour is pre-set and wherein said preset rule states that such f ul 1 -contour is selected f rom among said plural f ull-contour which has the largest full-contour data length and longer full-contour data is sequentially selected f rom among the remaining f ull-con-Lour data excluding both f ul 1-contour data lying ahead and at back of said longest fullcontour data.
203 42. A method for forming a font of a character or a symbol comprising:
inputting a plurality of coordinate data lying on a contour of a stroke specifying a dash of the character or the symbol; preparing data of the stroke including a plurality of input coordinate data, having one of the input coordinate data as coordinate data specifying the beginning point of the stroke or as coordinate data specifying the end point of the stroke; and forming a font of the character or the symbol including the data of one or more of the strokes.
43 - The font forming method as claimed in claim 42 wherein the initially entered coordinate data is the coordinate data specifying said beginning po-int.
44. The font forming method as claimed in claim 42 wherein the initially entered coordinate data is the coordinate data specifying said end point.
45. The font fo.rming method as claimed in claim 42wherein said font has data specifying a stroke sequence.
46. The font forming method as claimed in claim 42wherein data specifying the stroke sequence is formed in the stroke input sequence.
47. The font forming method as claimed in claim 42 comprising reading a character or a symbol by a scanner for storage in a memory; displaying the charac-,er or the symbol stored, in a 204 display device; and entering said plural coordinate data while having reference to the character or the symbol displayed on the display device. 48. The font forming method as claimed in claim 42comprising storing a pre-existing font in a memory; displaying a character or a symbol derived from the font stored in said memory on a display device; and entering said plural coordinate data while having reference to the character or the symbol displayed on the display device. 49. An apparatus for forming a font of a character or a symbol comprising: means for inputting a plurality of coordinate data lying on a contour of a stroke specifying a dash of the character or symbol; means for preparing data of the stroke including a plurality of input coordinate data, having one of the input coordinate data as coordinate data specifying the beglinning point of the stroke or as coordinate data specifying the end point of the stroke; and means for forming a font of the character or the symbol including the data of one or more of the strokes.
50. The font forming apparatus as claimed in claim 49 whereJ-n the initially entered coordinate data is the coordinate data specifying said beginning point.
51. The font forming apparatus as claimed in claim 49 where-Jn the last entered coordinate data is the coordinate data specifying 205 1 said end point. 52. The font forming apparatus as claimed in claim 49 wherein said font has data specifying the stroke sequence. 53. The font forming apparatus as claimed in claim 52 comprising stroke sequence setting means for forming said stroke sequence specifying data in accordance with the stroke data forming sequence in said stroke data forming'means. 54. A font forming apparatus as claimed in claim49 comprising: picture readout meansfor reading a character or a symbol for outputting image data; storage means for storing the image data from said picture readout means; and display means for displaying a picture of the character or the symbol based upon the image data stored in said storagemeans; said plural coordinate data being entered by said coordinate input means while having reference to the character or the symbol displayed on the display device. 55. A font forming apparatus as claimed in claim 49comprising: picture font storage means for storing a pre-existing font; and display means for displaying a picture of the character or the symbol based upon the font stored in said font storage means; said plural coordinate data being entered by said coordinate input means while having reference to the character or the symbol 206 1 / displayed on the display device.
56. A font of a character or a symbol wherein said character or the symbol is made up of one or more strokes each representing a dash of a writing brush and wherein said stroke data is made up of a plurality of coordinate data specifying the position of the stroke an the contour and coordinate data specifying the beginning point or the end point of the stroke.
57. The font as claimed in claim 56 wherein the coordinate data specifying said beginning point is the initially entered stroke coordinate data.
58. The font as claimed in claim 57 wherein the Coordinate data specifying said end point is the last entered stroke coordinate data.
59. The font as claimed in claim 58 having stroke sequence specif ying data.
60. The font as claimed in claim 59 wherein the stroke input sequence is the stroke sequence specifying data.
207 61. A method of forming a font of a character or symbol, the method being substantially as hereinbefore described with reference to the accompanying drawings.
62. Apparatus for forming a font of a character or symbol, the apparatus being substantially as hereinbefore described with reference to the accompanying drawings.
63. A font of a character or symbol, the font being substantially as hereinbefore described with reference to the accompanying drawings. 1 1 1 64. A method of representing a blurred figure on a display device, the method being substantially as hereinbefore described with reference to the accompanying drawings.
65. Apparatus for representing a blurred figure on a display device, the apparatus being substantially as hereinbefore described with reference to the accompanying drawings.
208
GB9922347A 1995-01-27 1996-01-26 Simulating a brush stroke with varying coverage Withdrawn GB2338160A (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
JP7012109A JPH08202336A (en) 1995-01-27 1995-01-27 Font forming method, font forming device and font
JP7012110A JPH08202893A (en) 1995-01-27 1995-01-27 Method and device for plotting blur
GB9601613A GB2299249A (en) 1995-01-27 1996-01-26 Method and apparatus for forming a font

Publications (2)

Publication Number Publication Date
GB9922347D0 GB9922347D0 (en) 1999-11-24
GB2338160A true GB2338160A (en) 1999-12-08

Family

ID=27268093

Family Applications (1)

Application Number Title Priority Date Filing Date
GB9922347A Withdrawn GB2338160A (en) 1995-01-27 1996-01-26 Simulating a brush stroke with varying coverage

Country Status (1)

Country Link
GB (1) GB2338160A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
AU729835B2 (en) * 1999-02-26 2001-02-08 Canon Kabushiki Kaisha Method and apparatus for modification of a natural image

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2256118A (en) * 1991-05-21 1992-11-25 Cambridge Animation Syst Image synthesis and processing
GB2297673A (en) * 1995-01-26 1996-08-07 Sony Corp Simulating ink spread in fibrous paper

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2256118A (en) * 1991-05-21 1992-11-25 Cambridge Animation Syst Image synthesis and processing
GB2297673A (en) * 1995-01-26 1996-08-07 Sony Corp Simulating ink spread in fibrous paper

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
AU729835B2 (en) * 1999-02-26 2001-02-08 Canon Kabushiki Kaisha Method and apparatus for modification of a natural image

Also Published As

Publication number Publication date
GB9922347D0 (en) 1999-11-24

Similar Documents

Publication Publication Date Title
US5611036A (en) Apparatus and method for defining the form and attributes of an object in an image
US5598182A (en) Image synthesis and processing
Steyvers Morphing techniques for manipulating face images
Fekete et al. TicTacToon: A paperless system for professional 2D animation
CN102509357B (en) Pencil sketch simulating and drawing system based on brush stroke
JP4404650B2 (en) Makeup simulation device, makeup simulation method, makeup simulation program
US5940081A (en) Method and apparatus for forming a font and the font produced method and apparatus for drawing a blurred figure
Anastacio et al. Modeling plant structures using concept sketches
CN111475667A (en) Mechanical arm regular script style calligraphy writing method based on handwritten Chinese characters
CN101697234A (en) Stroke segmentation modeling-based handwritten Chinese character Lishu beautifying method
US6437779B1 (en) Image processing method image processing unit and recording medium recording integrated shaping model data and image processing program to be used thereof
CN114298181A (en) Vector font generation method based on bimodal learning
Yang et al. Easy drawing: Generation of artistic chinese flower painting by stroke-based stylization
US5784301A (en) Method and apparatus for producing paper fiber structure data, and method and apparatus for drawing bled figure
GB2338160A (en) Simulating a brush stroke with varying coverage
Draper et al. A Gestural Interface to Free-Form Deformation.
JP3413829B2 (en) Drawing processing apparatus and method
JP3002972B2 (en) 3D image processing device
CN112232022A (en) Chinese character font and layout design method and system for dynamically adjusting parameters
Park et al. A feature‐based approach to facial expression cloning
Lee et al. Stroke Based Painterly Rendering with Mass Data through Auto Warping Generation.
JPH08202893A (en) Method and device for plotting blur
Ji et al. Artist-drawing inspired automatic sketch portrait rendering
CN110097644A (en) A kind of expression moving method, device, system and processor based on mixed reality
Todo et al. Stroke Transfer: Example-based Synthesis of Animatable Stroke Styles

Legal Events

Date Code Title Description
WAP Application withdrawn, taken to be withdrawn or refused ** after publication under section 16(1)