BACKGROUND OF THE INVENTION

1. Field of the Invention [0001]

The present invention relates to a method for evaluating the quality of a digital signal that has been read out from a storage medium and then decoded by a maximum likelihood decoding technique, and also relates to an apparatus for reading information from a storage medium and performing such quality evaluation on the read signal. [0002]

2. Description of the Related Art [0003]

Recently, various types of apparatuses (such as hard disk drive (HDD), optical disk drive and magnetooptical disk drive) for reading digital information from a storage medium have been used extensively in audiovisual appliances, personal computers and so on. [0004]

FIG. 1 is a block diagram showing a configuration for a part of a conventional optical disk drive [0005] 900. A light beam that has been reflected from an optical disk 1 is converted by an optical head 2 into a read signal. The read signal has its waveform shaped by a waveform equalizer 3 and then digitized by a comparator 4. The threshold value of the comparator 4 is normally subjected to a feedback control so that the output digital signals of the comparator 4 equals zero when integrated together.

In the optical disk drive [0006] 900, normally a phaselocked loop (PLL) circuit generates a clock signal that is synchronized with a read signal. A clock signal of that type is termed “a read clock signal”. As shown in FIG. 1, the PLL circuit includes a phase detector 5, a lowpass filter (LPF) 6 and a voltage controlled oscillator (VCO) 7. To generate the read clock signal, the phase detector 5 detects a difference in phase between the output digital signal of the comparator 4 and an output clock signal of the VCO 7. The phase difference detected is averaged by the LPF 6. In accordance with the output of the LPF 6, the control voltage of the VCO 7 is determined. In this manner, the oscillation frequency of the VCO 7 is subjected to a feedback control so that the phase difference output from the phase detector 5 always equals zero. Thus, the VCO 7 can output a clock signal that is synchronized with the read signal. By using a PLL circuit like this, even if the disk has some degree of eccentricity, for example, a clock signal can be extracted constantly so as to be synchronized with the read signal.

The read clock signal is used to determine whether the recorded code (i.e., digital information) is one or zero. More specifically, the digital information can be read out by determining whether or not each detection pulse of the comparator [0007] 4 falls within a window width defined by the read clock signal. As used herein, the “detection pulse” of the comparator 4 refers to a portion of the output digital signal of the comparator 4 that exceeds the predetermined threshold value.

However, the output detection pulse of the comparator [0008] 4 might deviate from the window width of the read clock signal due to various factors including intersymbol interference occurring in the read signal, the distortion of a recording mark, circuit noise and a control residual of the PLL. In that case, an error occurs. Such a time lag created between the detection pulse of the comparator 4 and the read clock signal is called a “jitter”.

In reading digital information by the technique described above, the quality (which is represented in terms of an error rate) of the read signal can be evaluated by using the distribution of jitter. The jitter distribution may be supposed to form a normal distribution having a mean of zero. In that case, the error rate Pj (σ/Tw) is given by
[0009] $\begin{array}{cc}\mathrm{Pj}\ue8a0\left(\sigma /\mathrm{Tw}\right)=2\ue89e\text{\hspace{1em}}\ue89e\mathrm{erfc}\ue89e\text{\hspace{1em}}\ue89e\left(\frac{\mathrm{Tw}/2}{\sigma}\right)& \left(1\right)\\ \mathrm{erfc}\ue8a0\left(z\right)=\frac{1}{\sqrt{2\ue89e\pi}}\ue89e{\int}_{z}^{\infty}\ue89e\mathrm{exp}\ue8a0\left(\frac{{u}^{2}}{2}\right)\ue89e\uf74cu& \left(2\right)\end{array}$

where σ is the standard deviation of the jitter distribution that is supposed to be a normal distribution and Tw is the window width. [0010]

FIG. 2 is a graph showing a relationship between the jitter and the bit error rate (BER). As can be seen from FIG. 2, as the standard deviation of the jitter increases, the BER also increases. The jitter of a read signal can be actually measured with a time interval analyzer (TIA). Accordingly, even if no errors have actually occurred, the quality of the signal can also be evaluated by the jitter standard deviation σ per the window width Tw. Thus, it is possible to predict the probability of occurrence of errors (which will be herein referred to as an “error probability”). For that reason, by measuring the standard deviation of the jitter, the performance of a given drive, a storage medium or an optical head can be checked and tested. Also, if the parameters of an equalizer are controlled in such a manner as to decrease the standard deviation of the jitter, then a read operation can be performed even more constantly. [0011]

In the technique described above, digital information is directly obtained from the output digital signal of the comparator [0012] 4. According to another known technique on the other hand, digital information may also be obtained by a maximum likelihood decoding method. Examples of known maximum likelihood decoding methods include a partial response maximum likelihood (PRML) method. In the PRML method, data is read or written from/on a storage medium having a high storage capacity with the potential occurrence of intersymbol interference fully taken into account. More specifically, a signal that has been read out from such a highcapacity storage medium is subjected, by a waveform equalizer, a digital filter and so on, to a partial response equalization so as to have a predetermined frequency characteristic. Then, the PR equalized and filtered signal is decoded into most likely (or most probable) digital data by a Viterbi decoding technique, for example. According to the PRML method, data can be decoded at a low error rate even from a read signal with a low signaltonoise ratio (SNR) or a read signal that is affected by the intersymbol interference relatively seriously.

In a maximum likelihood decoding method like this, data is decoded from a read signal by selecting a most probable state transition path. In general, a quantity representing the probability of a state transition that leads to a state Sn (where n is a state number) at a time k is defined by the following Equation (3):
[0013] $\begin{array}{cc}{L}^{\mathrm{Sn}}=\sum _{i=0}^{k}\ue89e{\left({y}_{i}{\mathrm{level}}_{v}\right)}^{2}& \left(3\right)\end{array}$

where y[0014] _{1 }is the actual value of the read signal (or digital sample data) at a time i and level_{v }is an expected ideal value of the read signal.

In a maximum likelihood decoding method, a state transition path having the minimum probability quantity as represented by Equation (3) is selected. Unlike the abovedescribed technique of decoding the data as one or zero by determining whether or not the detection pulse falls within the window width at each point in time k, a Euclidean distance of (y[0015] _{k}level_{v})^{2 }is obtained from the data that is sampled at each point in time k by reference to a read clock signal according to the maximum likelihood decoding method. Then, the data is decoded based on the Euclidean distance. Accordingly, the decoded result obtained by the maximum likelihood decoding method is also affected by a past sampled value y_{k }of a read signal.

In this maximum likelihood decoding method, even when two read signals have the same jitter standard deviation σ, errors may or may not have occurred in the read signals. For that reason, it is difficult to estimate the error rate of the decoded digital data, obtained by the maximum likelihood decoding method, by the jitter standard deviation σ of the read signal. Accordingly, an error rate estimating method (i.e., a signal quality evaluating method), which is more suitable to the maximum likelihood decoding method, needs to be used. [0016]

A method for evaluating the quality of a signal that has been decoded by the maximum likelihood decoding method is disclosed in Japanese LaidOpen Publication No. 1021651, for example. The apparatus disclosed in Japanese LaidOpen Publication No. 1021651 obtains a difference in likelihood between two state transition paths, having a minimum Euclidean distance between them, and then processes this difference by a statistical method, thereby evaluating the quality of the signal. [0017]

More specifically, to obtain a difference in likelihood between two paths that result in the same state at a time k, the sums of branch metrics of two survived paths that were regarded as most likely for two mutually different states at the previous time k−[0018] 1 are used. However, these sums of branch metrics at the time k−1 might be those of unwanted paths. For example, a path other than the path in question (i.e., a path having likelihood to be checked) may have been selected by mistake before the time k−1. Japanese LaidOpen Publication No. 1021651 does disclose a technique of selecting two paths having the minimum Euclidean distance between them and obtaining a difference in likelihood between these two paths. However, Japanese LaidOpen Publication No. 1021651 does not disclose any specific method for calculating the target likelihood values of these two paths with more certainty.
SUMMARY OF THE INVENTION

In order to overcome the problems described above, preferred embodiments of the present invention provide a method and apparatus for evaluating the quality of a read signal by using indices that are correlated with the error rate of digital data decoded by the maximum likelihood decoding method. [0019]

A preferred embodiment of the present invention provides a method for evaluating the quality of a read signal that has been decoded by a maximum likelihood decoding method. In this method, a most probable state transition path is preferably selected from a number n (where n is an integer equal to or greater than two) of state transition paths that represent n probable transitions from a first state S[0020] _{k−j }(where k is an integer equal to or greater than three and j is an integer equal to or greater than two) at a time k−j into a second state S_{k }at a time k. The method preferably includes the step of detecting predetermined combinations of the first and second states S_{k−j }and S_{k }that define the n probable state transition paths in a predetermined period j between the times k−j and k. The method preferably further includes the step of evaluating the reliability of the decoded signal, obtained in the predetermined period j, by using Pa−Pb . Pa and Pb are indices indicating the respective probabilities of state transition of first and second state transition paths in the predetermined period j. The first and second state transition paths are estimated to be the most probable and the second most probable, respectively, among the n probable state transition paths that are defined by the predetermined combinations detected.

In one preferred embodiment of the present invention, the step of evaluating the reliability preferably includes the steps of defining the index Pa by differences between expected values shown by the first state transition path and actual sample values in the predetermined period j, and defining the index Pb by differences between expected values shown by the second state transition path and the actual sample values in the predetermined period j. [0021]

In this particular preferred embodiment, the step of evaluating the reliability preferably includes the steps of obtaining the index Pa as a sum of squares of differences between the expected values l[0022] _{k−j}, . . . , l_{k−1 }and l_{k }shown by the first state transition path and the actual sample values y_{k−j}, . . . , y_{k−1 }and y_{k }in the predetermined period j and obtaining the index Pb as a sum of squares of differences between the expected values m_{k−j}, . . . , m_{k−1 }and m_{k }shown by the second state transition path and the actual sample values y_{k−j}, . . . , y_{k−1 }and Y_{k }in the predetermined period j.

In another preferred embodiment of the present invention, the number n is preferably two. [0023]

In still another preferred embodiment, a Euclidean distance between the first and second state transition paths is preferably a minimum value. [0024]

In yet another preferred embodiment, the method preferably further includes the step of detecting a variation in the reliability of the decoded signal by measuring Pa−Pb a number of times. [0025]

In this particular preferred embodiment, the step of detecting the variation in the reliability may include the step of deriving a standard deviation of a Pa−Pb distribution as the variation. [0026]

Alternatively, the step of detecting the variation in the reliability may include the step of deriving a standard deviation and an average of a Pa−Pb distribution as the variation. [0027]

As another alternative, the step of detecting the variation in the reliability may include the step of detecting a frequency of occurrence at which Pa−Pb exceeds a predetermined range. [0028]

In yet another preferred embodiment, the method may further include the step of decoding a read signal in which a recorded code has a minimum polarity inversion interval of two and which has been subjected to a PR (C0, C1, C0) equalization. [0029]

Alternatively, the method may further include the step of decoding a read signal in which a recorded code has a minimum polarity inversion interval of two and which has been subjected to a PR (C0, C1, C1, C0) equalization. [0030]

As another alternative, the method may further include the step of decoding a read signal in which a recorded code has a minimum polarity inversion interval of two and which has been subjected to a PR (C0, C1, C2, C1, C0) equalization. [0031]

In yet another preferred embodiment, the step of evaluating the reliability may include the step of obtaining Pa−Pb without calculating squares of the actual sample values. [0032]

Another preferred embodiment of the present invention provides an apparatus for reading information. The apparatus preferably includes gain controller, first waveform equalizer, read clock signal generator, A/D converter, maximum likelihood decoder and differential metric calculator. The gain controller preferably adjusts an amplitude value of a read signal. The first waveform equalizer preferably shapes the waveform of the read signal so that the read signal has a predetermined equalization characteristic. The read clock signal generator preferably generates a read clock signal that is synchronized with the read signal. The A/D converter preferably generates and outputs sampled data by sampling the read signal in response to the read clock signal. The maximum likelihood decoder preferably decodes the sampled data into most likely digital information. The differential metric calculator preferably obtains Pa−Pb . Pa and Pb are indices indicating respective probabilities of state transition of first and second state transition paths in a predetermined period. The first and second state transition paths are estimated by the maximum likelihood decoder to be the most probable and the second most probable, respectively. [0033]

In one preferred embodiment of the present invention, the apparatus preferably further includes a second waveform equalizer for shaping the waveform of the read signal differently from the first waveform equalizer so that the read signal has another predetermined equalization characteristic. In that case, the read clock signal is preferably generated from the read signal that has had its waveform shaped by the second waveform equalizer. [0034]

Other features, elements, processes, steps, characteristics and advantages of the present invention will become more apparent from the following detailed description of preferred embodiments of the present invention with reference to the attached drawings.[0035]
BRIEF DESCRIPTION OF THE DRAWINGS

FIG. 1 is a block diagram illustrating a configuration for a conventional optical disk drive. [0036]

FIG. 2 is a graph showing a relationship between the jitter and the bit error rate (BER). [0037]

FIG. 3 is a state transition diagram that is defined by the constraints, including a minimum polarity inversion interval of two and the use of a PR (1, 2, 2, 1) equalization technique, according to a preferred embodiment of the present invention. [0038]

FIG. 4 is a trellis diagram that is defined by the constraints, including the minimum polarity inversion interval of two and the use of the PR (1, 2, 2, 1) equalization technique, in the preferred embodiment of the present invention. [0039]

FIG. 5 is a diagram showing two possible state transition paths between states SO[0040] _{k }and SO_{k−5 }that are extracted from the trellis diagram shown in FIG. 4.

FIGS. 6A and 6B are graphs schematically showing the distributions of the reliability Pa−Pb of the decoded result. [0041]

FIG. 7 is a block diagram illustrating a configuration for an optical disk drive as an exemplary apparatus for evaluating the quality of a read signal according to a third specific preferred embodiment of the present invention. [0042]

FIG. 8 is a block diagram illustrating detailed configurations of the Viterbi circuit and differential metric analyzer of the optical disk drive shown in FIG. 7. [0043]

FIG. 9 is a diagram illustrating a detailed configuration of the path memory of the optical disk drive shown in FIG. 7. [0044]

FIG. 10 is a block diagram illustrating a configuration for another optical disk drive according to the third preferred embodiment. [0045]

FIG. 11 is a block diagram illustrating a configuration for still another optical disk drive according to the third preferred embodiment. [0046]

FIG. 12 is a block diagram illustrating a configuration for yet another optical disk drive according to the third preferred embodiment. [0047]

FIG. 13 is a block diagram illustrating a configuration for an optical disk drive according to a fourth specific preferred embodiment of the present invention. [0048]

FIG. 14 is a block diagram illustrating a configuration for another optical disk drive according to the fourth preferred embodiment. [0049]

FIG. 15 is a block diagram illustrating a configuration for still another optical disk drive according to the fourth preferred embodiment. [0050]

FIG. 16 is a graph showing a relationship between the PRML error index MLSA and the bit error rate (BER). [0051]

FIG. 17 is a state transition diagram that is defined by the constraints, including a minimum polarity inversion interval of two and the use of a PR (C0, C1, C0) equalization technique, according to another preferred embodiment of the present invention. [0052]

FIG. 18 is a trellis diagram that is defined by the constraints, including the minimum polarity inversion interval of two and the use of the PR (C0, C1, C0) equalization technique, in the preferred embodiment of the present invention. [0053]

FIG. 19 is a state transition diagram that is defined by the constraints, including a minimum polarity inversion interval of two and the use of a PR (C0, C1, C2, C1, C0) equalization technique, according to another preferred embodiment of the present invention. [0054]

FIG. 20 is a trellis diagram that is defined by the constraints, including the minimum polarity inversion interval of two and the use of the PR (C0, C1, C2, C1, C0) equalization technique, in the preferred embodiment of the present invention.[0055]
DETAILED DESCRIPTION OF PREFERRED EMBODIMENTS
Embodiment 1

Hereinafter, a method for evaluating the quality of a read signal and an apparatus for reading information according to preferred embodiments of the present invention will be described with reference to the accompanying drawings. [0056]

First, a read signal quality evaluating method according to a preferred embodiment of the present invention will be described. In the preferred embodiments to be described below, a code having a minimum polarity inversion interval of two, e.g., a code defined by a ([0057] 1, 7) RLL modulation method, is used as the recorded code. That is to say, any recorded code always has two or more consecutive zeros or ones. Also, a signal is supposed to be decoded by a PRML method in which the frequency characteristics of read and write systems substantially correspond to a PR (1, 2, 2, 1) equalization characteristic as a whole. Hereinafter, a specific decoding procedure will be described.

Suppose recorded codes (each being made up of zeros or ones) are denoted by: [0058]

Current recorded code: b[0059] _{k};

Recording code at the previous time: b[0060] _{k−1};

Recording code at the second last time: b[0061] _{k−2}; and

Recording code at the third last time: b[0062] _{k−3 }

An ideal value Level[0063] _{v }of a read signal that has been subjected to the PR (1, 2, 2, 1) equalization is given by:

Level_{v} =b _{k−3}+2b _{k−2}+2b _{k−1} +b _{k} (4)

where k is an integer representing the time and v is an integer of 0 to 6. In the PR (1, 2, 2, 1) equalization, there are seven ideal sample values (or expected values) Level[0064] _{v }(where 0≦v≦6) depending on the combination of the recorded codes.

Next, the state transitions of the recorded codes will be described. A state at the time k is represented by S (b
[0065] _{k−2}, b
_{k−1}, b
_{k}) and a state at the previous time k−
1 is represented by S (b
_{k−3}, b
_{k−2}, b
_{k−1}). The following Table 1 is a table of state transitions that is compiled by obtaining possible combinations of states at the times k−
1 and k. As described above, a modulation technique that defines the minimum inversion interval at two (i.e., at least two zeros or ones appear consecutively) is adopted in this preferred embodiment. Accordingly, the possible state transitions that a recorded code can make are limited to the following ten:
 TABLE 1 
 
 
 State  State  
 S (b_{k3}, b_{k2}, b_{k1})  S (b_{k2}, b_{k1}, b_{k})  
 at time k1  at time k  b_{k}/Level_{v} 
 
 S (0, 0, 0)  S (0, 0, 0)  0/0 
 S (0, 0, 0)  S (0, 0, 1)  1/1 
 S (0, 0, 1)  S (0, 1, 1)  1/3 
 S (0, 1, 1)  S (1, 1, 0)  0/4 
 S (0, 1, 1)  S (1, 1, 1)  1/5 
 S (1, 0, 0)  S (0, 0, 0)  0/1 
 S (1, 0, 0)  S (0, 0, 1)  1/2 
 S (1, 1, 0)  S (1, 0, 0)  0/3 
 S (1, 1, 1)  S (1, 1, 0)  0/5 
 S (1, 1, 1)  S (1, 1, 1)  1/6 
 

In the following description, the states S (0, 0, 0)[0066] _{k}, S (0, 0, 1)_{k}, S (0, 1, 1)_{k}, S (1, 1, 1)_{k}, S (1, 1, 0)_{k}, S (1, 0, 0)_{k }and so on at the time k will be identified by S0 _{k}, S1 _{k}, S2 ^{k}, S3 ^{k}, S4 ^{k}, S5 ^{k }and so on, respectively, for the sake of simplicity. The state transitions that may occur in the period between the time k−1 and the time k (i.e., a period of time corresponding to one cycle T of a read clock signal) are represented by the state transition diagram shown in FIG. 3. When the state transition diagram shown in FIG. 3 is expanded with respect to the time axis, the trellis diagram shown in FIG. 4 is obtained.

Look at the state S[0067] 0 _{k }at the time k and the state So_{k−5 }at a time k−5 shown in FIG. 5. In FIG. 5, two possible paths of state transitions that can occur between the states S0 _{k }and S0 ^{k−5 }are indicated by the bold arrows. One A of the two possible state transition paths goes through the states S0 _{k−5}. S0 ^{k−4}, S0 ^{k−3}, S0 ^{k−2}, S0 ^{k−1}, and S0 _{k }while the other possible state transition path B goes through the states S0 ^{k−5}, S1 _{k−4}, S2 ^{k−3}, S4 _{k−2}, S5 _{k−1 }and S0 ^{k}. In FIGS. 4 and 5, (recorded code/Level_{v}) is shown for each state transition. In this case, Level_{v }is supposed to be any value between −3 and 3. That is to say, the Level_{v }values of −3, −2, −1, 0, 1, 2 and 3 correspond to Level_{0}, Level_{1}, Level_{2}, Level_{3}, Level_{4}, Level_{5}, and Level_{6}, respectively.

In this manner, if the states at the times k−[0068] 5 and k are S0 ^{k−5 }and S0 ^{k}, respectively, it is estimated that the state transitions should have occurred along one of the two paths A and B. That is to say, if the decoded result obtained, i.e., the data that has been decoded in a period between the times k−7 and k, is (C_{k−7}, C_{k−6}, C_{k−5}, C_{k−4}, C_{k−3}, C_{k−2}, C_{k−1}, C_{k})=(0, 0, 0, x, x, 0, 0, 0) (where x is either zero or one), then the state transition path A or B is estimated to be the most probable one.

If the states S[0069] 0 _{k }and S0 _{k−5 }are detected at the times k and k−5, respectively (i.e., if the decoded result obtained is (0, 0, 0, x, x, 0, 0, 0)), then it is necessary to determine which is more probable, the path A or the path B. This decision can be made by comparing the sum of errors (or differences) between ideal sample values (i.e., expected values) shown by the path A and actual sample values with that of errors between ideal sample values (i.e., expected values) shown by the path B and the actual sample values. More specifically, for each of these two paths A and B, errors between expected values Level_{v }at the times k−4, k−3, k−2, k−1 and k and the actual values Y_{k−4}, Y_{k−3}, Y_{k−2}, Y_{k−1 }and Y_{k }of the read signal are obtained. The squares of these errors are summed up, thereby obtaining a sum of squared errors for each of the paths A and B. By comparing these sums of squared errors of the paths A and B with each other, it is possible to determine which of the two paths A and B is the more probable one.

In this case, the sum of squared errors between the expected values l[0070] _{k−4}, l_{k−3}, l_{k−2}, l_{k−1 }and l_{k }(i.e., 0, 0, 0, 0, 0) shown by the path A and the values Y_{k−4}, Y_{k−3}, Y_{k−2}, Y_{k−1 }and Y_{k }of the read signal at the times k−4, k−3, k−2, k−1 and k, respectively, is identified by Pa. On the other hand, the sum of squared errors between the expected values m_{k−4}, m_{k−3}, m_{k−2}, m_{k−1 }and m_{k }(i.e., 1, 3, 4, 3, 1) shown by the path B and the values Y_{k−4}, Y_{k−3}, Y_{k−2}, Y_{k−1}and Y_{k }of the read signal at the times k−4, k−3, k−2, k−1 and k, respectively, is identified by Pb. These sums of squared errors Pa and Pb are respectively given by the following Equations (5) and (6):

Pa=(y _{k−4}−0)^{2}+(y _{k−3}−0)^{2}+(y _{k−2}−0)^{2}+(y _{k−1}−0)^{2}+(y _{k}−0)^{2} (5)

Pb=(y _{k−4}−1)^{2}+(y _{k−3}−3)^{2}+(y _{k−2}−4)^{2}+(y _{k−1}−3)^{2}+(y _{k}−1)^{2} (6)

The sum of squared errors Pa obtained in this manner is an index indicating the probability of state transitions of the path A in the predetermined period between the times k−[0071] 5 and k. That is to say, the smaller the Pa value, the more probable the path A will be. On the other hand, the sum of squared errors Pb obtained in this manner is an index indicating the probability of state transitions of the path B in the predetermined period between the times k−5 and k. That is to say, the smaller the Pb value, the more probable the path B will be. Also, if the Pa or Pb value is zero, then the path A or B is estimated to be the most probable one.

Next, it will be described what the difference Pa−Pb between the Pa and Pb values means. A maximum likelihood decoder does not hesitate to choose the path A if Pa<<Pb or the path B if Pa>>Pb. However, if Pa=Pb, then either the path A or the path B may be chosen and the decoded result may or may not be true. Thus, the Pa−Pb value may be used as a measure of the reliability of the decoded result. That is to say, the greater the absolute value of Pa−Pb, the higher the reliability of the decoded result should be. On the other hand, the closer to zero the absolute value of Pa−Pb, the lower the reliability of the decoded result should be. [0072]

This Pa−Pb index indicating the reliability of the decoded result is used to evaluate the quality of the read signal. For that purpose, by obtaining the Pa−Pb values for a predetermined amount of time or a predetermined number of times based on the decoded results, a Pa−Pb distribution is obtained. FIGS. 6A and 6B schematically illustrate the Pa−Pb distributions. Specifically, FIG. 6A shows a Pa−Pb distribution where noise is superposed on the read signal. As shown in FIG. 6A, this distribution has two peaks. One of the two peaks corresponds to a frequency of occurrence that reaches a local maximum value when Pa=0. The other peak corresponds to a frequency of occurrence that reaches a local maximum value when Pb=0. The Pa−Pb value corresponding to the zero Pa value will be herein identified by −Pstd and the Pa−Pb value corresponding to the zero Pb value will be herein identified by Pstd. When Pstd is subtracted from the absolute value of Pa−Pb (i.e., when Pa−Pb −Pstd is calculated), the distribution shown in FIG. 6B is obtained. [0073]

By supposing this distribution to be a normal distribution, the standard deviation σ and the average Pave of the distribution are obtained. The standard deviation σ and average Pave of this distribution may be used to estimate a bit error rate. For example, if the estimated Pa−Pb  distribution is curved gently and defined by a function that may have a value of zero or less (i.e., unless the frequency of occurrence of Pa−Pb =0 is zero), decoding errors may be regarded as occurring at a frequency of occurrence that corresponds to the probability at which the function becomes zero or less. In that case, the error probability P ( σ, Pave) may be defined by the following Equation (7) using the standard deviation σ and the average Pave:[0074]

P(σ, Pave)=erfc(Pstd+Pave/σ) (7)

In this manner, the error rate of the digital decoded result obtained by a maximum likelihood decoding method can be estimated by using the average Pave and the standard deviation σ that have been derived from the distribution of Pa−Pb. In other words, the average Pave and the standard deviation σ may be used as indices to the quality of the read signal. In the example described above, the Pa−Pb  distribution is supposed to be a normal distribution. But if it is difficult to regard the Pa−Pb distribution as a normal distribution, then it is possible to count how many times the Pa−Pb values are equal to or smaller than a predetermined reference value instead of deriving the average Pave and the standard deviation σ as described above. The count obtained in this manner may be used as an index indicating the degree of variance of the Pa−Pb values. [0075]

According to the preferred embodiment described above, if a state transition has occurred from a first predetermined state (e.g., S[0076] 0 _{k−5}) into a second predetermined state (e.g., S0 _{k}) during a predetermined period, the absolute value of the difference Pa−Pb between the probabilities of two possible paths in the predetermined period is calculated, thereby evaluating the reliability of the decoded result. Furthermore, by measuring the Pa−Pb  values a number of times, a variance (or distribution) representing the degree of reliability Pa−Pb of the decoded result can be obtained. In this manner, the quality of the read signal can be evaluated (i.e., the bit error rate of the read signal can be estimated).

It should be noted that in evaluating the signal quality by such a method, a combination of states, between which a state transition should have occurred along one of the two paths having the highest error probabilities (i.e., two paths having a minimum Euclidean distance between them), is preferably selected, and the signal quality may be evaluated by reference to the absolute value Pa−Pb of the difference between the probabilities of these two paths. Hereinafter, this point will be described in detail. [0077]

In decoding a read signal in accordance with the abovedescribed state transition rule that requires the use of a minimum polarity inversion interval of two in combination with the PR (1, 2, 2, 1) equalization, not just the S
[0078] 0 _{k−5}→S
0 _{k }state transition but also fifteen other state transitions may occur along two paths in the period between the times k−
5 and k. The following Table 2 lists those sixteen state transitions (i.e., 16 combinations of states at the times k−
5 and k) with their respective possible Pa−Pb (or Pstd) values:
TABLE 2 


 Reliability PaPb  
State  of decoded result 
transition  If Pa = 0  If Pb = 0 

S0_{k5 } S0_{k}  −36  +36 
S0_{k5 }→ S1_{k}  −36  +36 
S0_{k4 }→ S4_{k}  −10  +10 
S0_{k4 }→ S3_{k}  −10  +10 
S2_{k4 }→ S0_{k}  −10  +10 
S2_{k4 }→ S1_{k}  −10  +10 
S2_{k5 }→ S4_{k}  −36  +36 
S2_{k5 }→ S3_{k}  −36  +36 
S5_{k5 }→ S0_{k}  −36  +36 
S5_{k5 }→ S1_{k}  −36  +36 
S5_{k4 }→ S4_{k}  −10  +10 
S5_{k4 }→ S3_{k}  −10  +10 
S3_{k4 }→ S0_{k}  −10  +10 
S3_{k4 }→ S1_{k}  −10  +10 
S3_{k5 }→ S4_{k}  −36  +36 
S3_{k5 }→ S3_{k}  −36  +36 


The reliabilities Pa−Pb of the sixteen decoded results are given by the following Equations (8.1) through (8.16):[0079]

If (c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(0, 0, 0, x, x, 0, 0, 0), then Pa−Pb=(A _{k−4} −B _{k−4})+(A _{k−3} −D _{k−3})+(A _{k−2} −E _{k−2})+(A _{k−1} −D _{k−1})+(A _{k} −B _{k}) (8.1)

If (c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=

(0, 0, 0, [0080] x, x, 0, 0, 1), then Pa−Pb=(A _{k−4} −B _{k−4})+(A _{k−3} −D _{k−3})+(A _{k−2} −E _{k−2})+(A _{k−1} −D _{k−1})+(B _{k} −C _{k}); (8.2)

If (c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(0, 0, 0, x, 1, 1, 0), then Pa−Pb=(A_{k−3} −B _{k−3})+(B _{k−2} −D _{k−2})+(D _{k−1} −F _{k−1})+(E _{k} −F _{k}); (8.3)

If (c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(0, 0, 0, x, 1, 1, 1), then Pa−Pb=(A _{k−3} −B _{k−3})+(B _{k−2} −D _{k−2})+(D _{k−1} F _{k−1})+(F_{k} −G _{k}); (8.4)

If (c_{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(0, 1, 1, x, 0, 0, 0), then Pa−Pb=(E _{k−3} −F _{k−3})+(D _{k−2} −F _{k−2})+(B _{k−1} −D _{k−1})+(A _{k} −B _{k}); (8.5)

If (c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(0, 1, 1, x, 0, 0, 1), then Pa−Pb=(E _{k−3} −F _{k−3})+(D _{k−2} −F _{k−2})+(B _{k−1} −D _{k−1})+(B _{k} −C _{k}); (8.6)

If (c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(0, 1, 1, x, x, 1, 1, 0), then Pa−Pb=(E _{k−4} −F _{k−4})+(D _{k−3} −G _{k−3})+(C _{k−2} −G _{k−2})+(D _{k−1} −G _{k−1})+(E _{k} −F _{k}); (8.7)

If (c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(0, 1, 1, x, x, 1, 1, 1), then Pa−Pb=(E _{k−4} −F _{k−4})+(D _{k−3} −G _{k−3})+(C _{k−2} −G _{k−2})+(D _{k−1} −G _{k−1})+(F _{k} −G _{k}); (8.8)

If (c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(1, 0, 0, x, x, 0, 0, 0), then Pa−Pb=(B _{k−4} −C _{k−4})+(A _{k−3} −D _{k−3})+(A _{k−2} −E _{k−2})+(A _{k−1} −D _{k−1})+(A _{k} −B _{k}); (8.9)

If (c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(1, 0, 0, x, x, 0, 0, 1), then Pa−Pb=(B _{k−4} −C _{k−4})+(A _{k−3} −D _{k−3})+(A _{k−2} −E _{k−2})+(A _{k−1} −D _{k−1})+(B _{k} −C _{k}); (8.10)

If (c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(1, 0, 0, x, 1, 1, 0), then Pa−Pb=(B _{k−3} −C _{k−3})+(B _{k−2} −D _{k−2})+(D _{k−1} −F _{k−1})+(E _{k} −F _{k}); (8.11)

If (c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(1, 0, 0, x, 1, 1, 1), then Pa−Pb=(B _{k−3} −C _{k−3})+(B _{k−2} −D _{k−2})+(D _{k−1} −F _{k−1})+(F_{k} −G _{k}); (8.12)

If (c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(1, 1, 1, x, 0, 0, 0), then Pa−Pb=(F _{k−3} −G _{k−3})+(D _{k−2} −F _{k−2})+(B _{k−1} −D _{k−1})+(A _{k} −B _{k}); (8.13)

If (c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(1, 1, 1, x, 0, 0, 1), then Pa−Pb=(F _{k−3} −G _{k−3})+(D _{k−2} −F _{k−2})+(B _{k−1} −D _{k−1})+(B _{k} −C _{k}); (8.14)

If (c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(1, 1, 1, x, x, 1, 1, 0), then Pa−Pb=(F _{k−4} −G _{k−4})+(D _{k−3} −G _{k−3})+(c_{k−2}−G_{k−2})+(D _{k−1} −G _{k−1})+(E _{k} −F _{k}); (8.15)

and [0081]

If (c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(1, 1, 1, x, x, 1, 1, 1), then Pa−Pb=(F _{k−4} −G _{k−4})+(D _{k−3} −G _{k−3})+(C _{k−2 } −G _{k−2})+(D _{k−1} −G _{k−1})+(F _{k} −G _{k}). (8.16)

where A[0082] _{k}=(y_{k}−0)^{2}, B_{k}=(y_{k}−1)^{2}, C_{k}=(y_{k}−2)^{2}, D_{k}=(y_{k}−3)^{2}, E_{k}=(y_{k}−4)^{2}, F_{k}=(y_{k}−5)^{2 }and G_{k}=(y_{k}−6)^{2}.

These Equations (8.1) through (8.16) may be classified by the Pstd value into the following two group of Equations (9.1) through (9.8) (where Pstd=10) and (10.1) through (10.8) (where Pstd=36): [0083]

If ([0084] c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})32 (0, 0, 0, x, 1, 1, 0), then Pa−Pb=(A _{k−3} −B _{k−3})+(B _{k−2} −D _{k−2})+(D _{k−1} −F _{k−1})+(E _{k} −F _{k}); (9.1)

If (c _{k−6} , c _{k−5} , c _{k−41} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(0, 0, 0, x, 1, 1, 1), then Pa−Pb=(A _{k−3} −B _{k−3})+(B _{k−2} −D _{k−2})+(D _{k−1} −F _{k−1})+(F _{k} −G _{k}); (9.2)

If (c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(0, 1, 1, x, 0, 0, 0), then Pa−Pb=(E _{k−3} −F _{k−3})+(D _{k−2} −F _{k−2})+(B _{k−1} −D _{k−1})+(A _{k} −B _{k}); (9.3)

If (c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(0, 1, 1, x, 0, 0, 1), then Pa−Pb=(E _{k−3} −F _{k−3})+(D _{k−2} −F _{k−2})+(B _{k−1} −D _{k−1})+(B _{k} −C _{k}); (9.4)

If (c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(1, 0, 0, x, 1, 1, 0), then Pa−Pb=(B _{k−3} −C _{k−3})+(B _{k−2} −D _{k−2})+(D _{k−1} −F _{k−1})+(E _{k} −F _{k}); (9.5)

If (c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(1, 0, 0, x, 1, 1, 1), then Pa−Pb=(B _{k−3} −C _{k−3})+(B _{k−2} −D _{k−2})+(D _{k−1} −F _{k−1})+(F _{k} −G _{k}); (9.6)

If (c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(1, 1, 1, x, 0, 0, 0), then Pa−Pb=(F _{k−3} −G _{k−3})+(D _{k−2} −F _{k−2})+(B _{k−1} −D _{k−1})+(A _{k} −B _{k}); (9.7)

and[0085]

If (c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(1, 1, 1, x, 0, 0, 1), then Pa−Pb=(F _{k−3} −G _{k−3})+(D _{k−2} −F _{k−2})+(B _{k−1} −D _{k−1})+(B_{k} −C _{k}) (9.8)

If (c _{k−7} , C _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(0, 0, 0, x, x, 0, 0, 0), then Pa−Pb=(A _{k−4}−B_{k−4})+(A _{k−3} −D _{k−3})+(A _{k−2} −E _{k−2})+(A _{k−1} −D _{k−1})+(A _{k} −B _{k}); (10.1)

If (c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(0, 0, 0, x, x, 0, 0, 1), then Pa−Pb=(A _{k−4} −B _{k−4})+(A _{k−3} −D _{k−3})+(A _{k−2 } −E _{k−2})+(A _{k−1} −D _{k−1})+(B _{k} −C _{k}) (10.2)

If (c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(0, 1, 1, x, x, 1, 1, 0), then Pa−Pb=(E _{k−4} −F _{k−4})+(D _{k−3} −G _{k−3})+(C _{k−2 } −G _{k−2})+(D _{k−1} −G _{k−1})+(E _{k} −F _{k}); (10.3)

If (c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(0, 1, 1, x, x, 1, 1, 1), then Pa−Pb=(E _{k−4} −F _{k−4})+(D _{k−3} −G _{k−3})+(C _{k−2 } −G _{k−2})+(D _{k−1} −G _{k−1})+(F _{k} −G _{k}); (10.4)

If (c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(1, 0, 0, x, x, 0, 0, 0), then Pa−Pb=(B _{k−4} −C _{k−4})+(A _{k−3} −D _{k−3})+(A _{k−2} −E _{k−2})+(A _{k−1} −D _{k−1})+(A _{k} −B _{k}) (10.5)

If (c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(1, 0, 0, x, x, 0, 0, 1), then Pa−Pb=(B _{k−4} −C _{k−4})+(A _{k−3} −D _{k−3})+(A _{k−2 } −E _{k−2})+(A _{k−1} −D _{k−1})+(B _{k} −C _{k}); (10.6)

If (c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(1, 1, 1, x, x, 1, 1, 0), then Pa−Pb=(F _{k−4} −G _{k−4})+(D _{k−3} −G _{k−3})+(C_{k−2} −G _{k−2})+(D _{k−1} −G _{k−1})+(E _{k} −F _{k}); (10.7)

and[0086]

If (c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(11, 1, x, x, 1, 1, 1), then Pa−Pb=(F _{k−4} −G _{k−4})+(D _{k−3} −G _{k−3})+(C_{k−2} −G _{k−2})+(D _{k−1} −G _{k−1})+(F _{k} −G _{k}) (10.8)

Hereinafter, it will be described how to obtain error rate indices to each of these possible state transitions. As for the eight state transitions having a Pstd value of 10, Pa−Pb values satisfying Equations (9.1) through (9.8) are obtained from the most likely decoded results c
[0087] _{k }and the standard deviation a σ
_{10 }and average Pave
_{10 }are obtained from their distribution. As for the other eight state transitions having a Pstd value of 36 on the other hand, Pa−Pb values satisfying Equations (10.1) through (10.8) are obtained from the most likely decoded results c
_{k }and the standard deviation a σ
_{36 }and average Pave
_{36 }are obtained from their distribution. Supposing these distributions are both normal distributions, the error probabilities P
_{10 }and P
_{36 }are respectively given by the following Equations (11) and (12):
$\begin{array}{cc}{P}_{10}\ue8a0\left({\sigma}_{10},{\mathrm{Pave}}_{10}\right)=\mathrm{erfc}\ue8a0\left(\frac{10+{\mathrm{Pave}}_{10}}{{\sigma}_{10}}\right)& \left(11\right)\\ {P}_{36}\ue8a0\left({\sigma}_{36},{\mathrm{Pave}}_{36}\right)=\mathrm{erfc}\ue8a0\left(\frac{36+{\mathrm{Pave}}_{36}}{{\sigma}_{36}}\right)& \left(12\right)\end{array}$

That is to say, the error rate can be estimated for each pattern of the most likely decoded results. Thus, the standard deviation a σ[0088] _{10 }and average Pave_{10 }or the standard deviation a σ_{36 }and average Pave_{36 }may be used as indices to the quality of the read signal.

If the state transition pattern detection range is expanded by one time period (i.e., when combination patterns of state transitions, each of which should have occurred along one of two paths, are detected in the period between the time k
[0089] 6 and the time k), the eight patterns shown in the following Table 3 may be further detected:
TABLE 3 


 Reliability PaPb  
State  of decoded result 
transition  If Pa = 0  If Pb = 0 

S0_{k6 }→ S0_{k}  −12  +12 
S0_{k6 }→ S1_{k}  −12  +12 
S2_{k6 }→ S3_{k}  −12  +12 
S2_{k6 }→ S4_{k}  −12  +12 
S5_{k6 }→ S0_{k}  −12  +12 
S5_{k6 }→ S1_{k}  −12  +12 
S3_{k6 }→ S3_{k}  −12  +12 
S3_{k6 }→ S4_{k}  −12  +12 


Just like the Equations (11) and (12), the error probability P
[0090] _{12 }of each of the eight patterns shown in Table 3 is given by the following Equation (13):
$\begin{array}{cc}{P}_{12}\ue8a0\left({\sigma}_{12},{\mathrm{Pave}}_{12}\right)=\mathrm{erfc}\ue8a0\left(\frac{12+{\mathrm{Pave}}_{12}}{{\sigma}_{12}}\right)& \left(13\right)\end{array}$

The point is that the reliability Pa−Pb may be used effectively as an index to the quality of the read signal by detecting only state transition patterns having relatively high error probabilities (or error rates). That is to say, an index correlated with the error rate can be obtained without detecting all state transition patterns. [0091]

As used herein, the “state transition patterns having relatively high error probabilities” refer to state transition patterns of which the maximum value of the reliability values Pa−Pb is relatively small (i.e., patterns having the smallest Euclidean distance, or the absolute distance between the paths A and B). In this example, the “state transition patterns having relatively high error probabilities” are the eight patterns shown in Table 2 in which Pa−Pb=±10 when Pa=0 or Pb=10. [0092]

If white noise prevails in the noise included in the read signal, it is expected that an inequality P[0093] _{10}>P_{12}>>P_{36 }is satisfied. Among these three error probabilities, only P_{10 }means a shift error of one bit, while the other two P_{12 }and P_{36 }mean a shift error of two or more bits. Generally speaking, almost all error patterns are found to be onebit shift errors after PRML processing has been performed. Accordingly, the error rate of the read signal may be appropriately estimated by Equation (11) that defines the error probability P_{10}. In this manner, the quality of the read signal may be evaluated by detecting patterns representing predetermined state transitions in a predetermined period and by using, as indices, the standard deviation a σ_{10 }and average Pave_{10 }of the Pa−Pb−Pstd distribution of the state transition patterns detected.

As described above, the error rate may be estimated by using the standard deviation a σ
[0094] _{10}. Alternatively, a maximum likelihood sequence amplitude (MLSA), which is an error index for use in PRML processing (which will be herein simply referred to as an “MLSA index”), may also be used as an index to the signal quality (or error rate). The MLSA index is defined by the following Equation (14):
$\begin{array}{cc}M=\frac{{\sigma}_{10}}{2\xb7{d}_{\mathrm{min}}^{2}}\ue89e\text{\hspace{1em}}\left[\%\right]& \left(14\right)\end{array}$

where d[0095] ^{2} _{min }is the square of the minimum Euclidean distance between two possible paths. In the combination of modulation code and PRML method as adopted in this preferred embodiment, d^{2} _{min}=10. This MLSA index is obtained by Equation (14) on the supposition that the average Pave_{10 }used in Equation (11) is zero (i.e., while leaving the average Pave_{10 }out of consideration). This is because the average Pave_{10 }is typically approximately zero and normally does not constitute a major factor of decreasing the correlation between the index and the error rate.

FIG. 16 shows a relationship between the MLSA index as defined by Equation (14) and a bit error rate BER as derived by Equation (11). It can be seen that just like the jittererror rate relationship shown in FIG. 2, as the MLSA index increases, the error rate increases. That is to say, it can be seen that the error rate to be obtained after the PRML processing may be estimated by using the MLSA index. [0096]

In the specific preferred embodiment described above, a PR (1, 2, 2, 1) equalization technique is used as an exemplary (C0, C1, C1, C0) equalization technique (where C0 and C1 are arbitrary positive integers). However, even if any other (C0, C1, C1, C0) equalization technique (where C0 and C1 are arbitrary positive integers) is adopted, an index correlated with the error rate can also be obtained through a similar procedure. [0097]

Hereinafter, another specific preferred embodiment of the present invention will be described. In the following specific preferred embodiment, a recorded code having a minimum polarity inversion interval of two is used as in the preferred embodiment described above. However, unlike the preferred embodiment described above, a PR (C0, C1, C0) (where C0 and C1 are arbitrary positive integers) equalization technique (e.g., PR (1, 2, 1) equalization) is applied to the following preferred embodiment. [0098]

Suppose recorded codes (each being made up of zeros or ones) are denoted by: [0099]

Current recorded code: b[0100] _{k};

Recording code at the previous time: b[0101] _{k−1}; and

Recording code at the second last time: b[0102] _{k−2 }

An ideal value Level, of a read signal that has been subjected to the PR (C0, C1, C0) equalization is given by the following Equation (15):[0103]

Level_{v} =C0×b _{k−2} +C1×b _{k−1} +C0×b _{k} (15)

where k is an integer representing a time and v is an integer of 0 to 3. Supposing a state at a time k is identified by S (b
[0104] _{k−1}, b
_{k}), a table of state transitions such as the following Table 4 is obtained:
 TABLE 4 
 
 
 State S (b_{k2}, b_{k1})  State S (b_{k1}, b_{k})  
 at time k1  at time k  b_{k}/LEVEL_{v} 
 
 S (0, 0)  S (0, 0)  0/0 
 S (0, 0)  S (0, 1)  1/C0 
 S (0, 1)  S (1, 1)  1/C0 + C1 
 S (1, 1)  S (1, 0)  0/C1 + C0 
 S (1, 1)  S (1, 1)  1/C0 + C1 + C0 
 S (1, 0)  S (0, 0)  0/C0 
 

In the following description, the states S (0, 0)[0105] _{k,}S (0, 1)_{k}, S (1, 1)_{k}, S (1, 0)_{k }and so on at the time k will be identified by S0 _{k}, S1 _{k}, S2 _{k}, S3 ^{k }and so on, respectively, for the sake of simplicity. The state transitions that may occur in the period between the time k−1 and the time k (i.e., a period of time corresponding to one cycle T of a read clock signal) are represented by the state transition diagram shown in FIG. 17. When the state transition diagram shown in FIG. 17 is expanded with respect to the time axis, the trellis diagram shown in FIG. 18 is obtained.

In this preferred embodiment, each recorded code has a minimum polarity inversion interval of two and the PR (C0, C1, C0) equalization technique is used. Under these conditions, there are six possible state transition patterns (i.e., possible combinations of states) as for state transitions occurring from a predetermined state at a time into another predetermined state at a different time along two paths (i.e., paths A and B) as shown in the following Table 5:
[0106] TABLE 5 


 Recording code  Recording code 
State  (b_{k1}, . . . , b_{k}) of  (b_{k1}, . . . , b_{k}) of 
transition  path A  path B 

S0_{k3 }→ S2_{k}  (0, 0, 0, 1, 1)  (0, 0, 1, 1, 1) 
S2_{k3 }→ S0_{k}  (1, 1, 0, 0, 0)  (1, 1, 1, 0, 0) 
S0_{k3 }→ S0_{k}  (0, 0, 0, 0, 0, 0)  (0, 0, 1, 1, 0, 0) 
S2_{k3 }→ S2_{k}  (1, 1, 0, 0, 1, 1)  (1, 1, 1, 1, 1, 1) 
S0_{k4 }→ S0_{k}  (0, 0, 0, 1, 1, 0, 0)  (0, 0, 1, 1, 0, 0, 0) 
S2_{k4 }→ S2_{k}  (1, 1, 0, 0, 0, 1, 1)  (1, 1, 1, 0, 0, 1, 1) 


In this case, it is determined which of the two paths A and B has the higher probability. This decision can be made by comparing the sum of errors between ideal sample values (i.e., expected values) shown by the path A and actual sample values with the sum of errors between associated ideal sample values (i.e., expected values) shown by the path B and the actual sample values. [0107]

For example, suppose the state transition S[0108] 0 _{k−3}→S2 _{k }should be estimated. As for this state transition, no matter whether the path A (S0 _{k−3}, S0 _{k−2}, S1 _{k−1}, S2 ^{k}) or the path B (S0 _{k−3}, S1 _{k−2}, S2 _{k−1}, S2 ^{k}) has been taken, the state at the time k−3 is S0 _{k−3 }and the state at the time k is S2 _{k}. In this case, it is by using the sum of squared errors between the expected values and the actual values y_{k−2}, y_{k−1 }and y_{k }of the read signal in the period between the times k−2 and k that it is determined which of the two paths A and B has the higher probability. As in the preferred embodiment described above, the sum of squared errors between the expected values shown by the path A and the actual values y_{k−2}, y_{k−1 }and y_{k }of the read signal in the period between the times k−2 and k is identified by Pa. On the other hand, the sum of squared errors between the expected values shown by the path B and the actual values y_{k−2}, y_{k−1}, and y_{k }of the read signal in the period between the times k−2 and k is identified by Pb. These sums of squared errors Pa and Pb (i.e., the error probabilities) are respectively given by the following Equations (16) and (17):

Pa=(y _{k−2}−0)^{2}+(y _{k−1} −C0)^{2}+(y _{k}−(C0+C1))^{2} (16)

Pb=(y _{k−2} −C0)^{2}+(y _{k−1}−(C0+C1))^{2}+(y _{k}−(2×C0+C1))^{2} (17)

In this case, if Pa<<Pb, then the path A is estimated to be the more probable one. On the other hand, if Pa>>Pb, then the path B is estimated to be the more probable one. That is to say, even when a recorded code having a minimum polarity inversion interval of two is combined with the PR (C0, C1, C0) equalization technique, the reliability of the decoded result can also be evaluated by Pa−Pb. Also, the quality of the read signal can be evaluated (or the error rate can be estimated) based on the Pa−Pb distribution. [0109]

Furthermore, suppose white noise has been superposed on the transmission line. In that case, a state transition having the highest error probability should have a minimum Euclidean distance between the paths A and B. The two state transition patterns shown in the following Table 6 should have the minimum Euclidean distance between their two paths:
[0110] TABLE 6 


 Recording code  Recording code 
State  (b_{k1}, . . . , b_{k})  (b_{k1}, . . . , b_{k}) 
transition  of path A  of path B 

S0_{k3 }→ S2_{k}  (0, 0, 0, 1, 1)  (0, 0, 1, 1, 1) 
S2_{k3 }→ S0_{k}  (1, 1, 0, 0, 0)  (1, 1, 1, 0, 0) 


The reliabilities Pa−Pb of the two state transition patterns shown in Table 6 are given by the following Equations (18.1) and (18.2):[0111]

If (c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(0, 0, x, 1, 1), then Pa−Pb=(AA _{k−2} −BB _{k−2})+(BB _{k−1} −CC _{k−1})+(CC_{k} −DD _{k}); (18.1)

and[0112]

If (c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(1, 1, x, 0, 0), then Pa−Pb=(CC _{k−2} −DD _{k−2})+(BB_{k−1} −CC _{k−})+(AA _{k} −BB _{k}) (18.2)

where c[0113] _{k }is the decoded result, k is an integer, and AA_{k}, BB_{k}, CC_{k }and DD_{k }are given by:

AA _{k}=(y _{k}−0)^{2},

BB _{k}=(y _{k} −C0)^{2},

CC _{k}=(y _{k}−(C0+C1))^{2}

and[0114]

DD _{k}=(y _{k}−(2×C0+C1))^{2}

Pa−Pb −(2×C0
[0115] ^{2}+C1
^{2}) that satisfies Equations (18.1) and (18.2) is derived from the decoded result c
_{k }and the standard deviation σ and average Pave are obtained from its distribution. Supposing that distribution is a normal distribution, the error probability is given by the following Equation (19):
$\begin{array}{cc}P\ue8a0\left(\sigma ,\mathrm{Pave}\right)=\mathrm{erfc}\ue8a0\left(\frac{\mathrm{Pave}}{\sigma}\right)& \left(19\right)\end{array}$

Thus, the standard deviation σ and average Pave may be used to estimate the error rate of the read signal or evaluate the quality of the read signal. [0116]

As described above, even when a recorded code having a minimum polarity inversion interval of two is used in combination with the PR (C0, C1, C0) equalization, the quality of the read signal can also be evaluated by the difference in probability Pa−Pb between the two paths of state transitions occurring in a predetermined period. [0117]

Hereinafter, another specific preferred embodiment of the present invention will be described. In the following specific preferred embodiment, a recorded code having a minimum polarity inversion interval of two is used as in the preferred embodiments described above. However, unlike the preferred embodiments described above, a PR (C0, C1, C2, C1, C0) (where C0, C1 and C2 are arbitrary positive integers) equalization technique is applied to the following preferred embodiment. [0118]

Suppose recorded codes are denoted by: [0119]

Current recorded code: b[0120] _{k};

Recording code at the previous time: b[0121] _{k−1};

Recording code at the second last time: b[0122] _{k−2};

Recording code at the third last time: b[0123] _{k−3}; and

Recording code at the fourth last time: b[0124] _{k−4 }

An ideal value Level[0125] _{v }of a read signal that has been subjected to the PR (C0, C1, C2, C1, C0) equalization is given by the following Equation (20):

Level_{v} =C0×b _{k−4} +C1×b _{k−3} +C2×b _{k−2} +C1×b _{k−1} +C0×b _{k} (20)

where k is an integer representing a time and v is an integer of 0 to 8. Supposing a state at a time k is identified by S (b
[0126] _{k−3}, b
_{k−2}, b
_{k−1}, b
_{k}), a table of state transitions such as the following Table 7 is obtained:
 TABLE 7 
 
 
 State  State  
 S (b_{k4}, b_{k3},  S (b_{k3}, b_{k2},  
 b_{k2}, b_{k1})  b_{k1}, b_{k})  
 at time k1  at time k  b_{k}/LEVEL_{v} 
 
 S (0, 0, 0, 0)  S (0, 0, 0, 0)  0/0 
 S (0, 0, 0, 0)  S (0, 0, 0, 1)  1/C0 
 S (0, 0, 0, 1)  S (0, 0, 1, 1)  1/C0 + C1 
 S (0, 0, 1, 1)  S (0, 1, 1, 0)  0/C1 + C2 
 S (0, 0, 1, 1)  S (0, 1, 1, 1)  1/C0 + C1 + C2 
 S (0, 1, 1, 0)  S (1, 1, 0, 0)  0/C1 + C2 
 S (0, 1, 1, 1)  S (1, 1, 1, 0)  0/2*C1 + C2 
 S (0, 1, 1, 1)  S (1, 1, 1, 1)  1/C0 + 2*C1 + C2 
 S (1, 0, 0, 0)  S (0, 0, 0, 0)  0/C0 
 S (1, 0, 0, 0)  S (0, 0, 0, 1)  1/2*C0 
 S (1, 0, 0, 1)  S (0, 0, 1, 1)  1/2*C0 + C1 
 S (1, 1, 0, 0)  S (1, 0, 0, 0)  0/C0 + C1 
 S (1, 1, 0, 0)  S (1, 0, 0, 1)  1/2*C0 + C1 
 S (1, 1, 1, 0)  S (1, 1, 0, 0)  0/C0 + C1 + C2 
 S (1, 1, 1, 1)  S (1, 1, 1, 0)  0/C0 + 2*C1 + C2 
 S (1, 1, 1, 1)  S (1, 1, 1, 1)  1/2*C0 + 2*C1 + C2 
 

In the following description, the states S (0, 0, 0, 0)[0127] _{k}, S (0, 0, 0, 1)_{k}, S (0, 0, 1, 1)_{k}, S (0, 1, 1)_{k}, S (1, 1, 1, 1 )_{k}, S (1, 1, 1, 0)_{k}, S (1, 1, 0, 0)_{k}, S (1, 0, 0, 0)_{k}, S (1, 0, 0, 1)_{k}, S (0. 1, 1. 0)_{k }and so on at the time k will be identified by S0 _{k}, S1 _{k}, S2 _{k}, S3 _{k}, S4 _{k}, S5 _{k}, S6 _{k}, S7 _{k}, S8 _{k}, S9 _{k }and so on, respectively, for the sake of simplicity. The state transitions that may occur in the period between the time k−1 and the time k (i.e., a period of time corresponding to one cycle T of a read clock signal) are represented by the state transition diagram shown in FIG. 19. When the state transition diagram shown in FIG. 19 is expanded with respect to the time axis, the trellis diagram shown in FIG. 20 is obtained.

In this preferred embodiment, each recorded code has a minimum polarity inversion interval of two and the PR (C0, C1, C2, C1, C0) equalization technique is used. Under these conditions, there are 90 possible state transition patterns (i.e., possible combinations of states) for state transitions occurring from a predetermined state at a time into another predetermined state at a different time along two paths (i.e., paths A and B) as shown in the following Table 8:
[0128] TABLE 8 


 Recording code  Recording code 
State  (b_{k1}, . . . b_{k})  (b_{k1}, . . . , b_{k}) 
transition  of path A  of path B 

S0_{k5 }→ S6_{k}  (0, 0, 0, 0, 0, 1, 1, 0, 0)  (0, 0, 0, 0, 1, 1, 1, 0, 0) 
S0_{k5 }→ S5_{k}  (0, 0, 0, 0, 0, 1, 1, 1, 0)  (0, 0, 0, 0, 1, 1, 1, 1, 0) 
S0_{k5 }→ S4_{k}  (0, 0, 0, 0, 0, 1, 1, 1, 1)  (0, 0, 0, 0, 1, 1, 1, 1, 1) 
S2_{k5 }→ S0_{k}  (0, 0, 1, 1, 0, 0, 0, 0, 0)  (0, 0, 1, 1, 1, 0, 0, 0, 0) 
S2_{k5 }→ S1_{k}  (0, 0, 1, 1, 0, 0, 0, 0, 1)  (0, 0, 1, 1, 1, 0, 0, 0, 1) 
S2_{k5 }→ S2_{k}  (0, 0, 1, 1, 0, 0, 0, 1, 1)  (0, 0, 1, 1, 1, 0, 0, 1, 1) 
S3_{k5 }→ S0_{k}  (0, 1, 1, 1, 0, 0, 0, 0, 0)  (0, 1, 1, 1, 1, 0, 0, 0, 0) 
S3_{k5 }→ S1_{k}  (0, 1, 1, 1, 0, 0, 0, 0, 1)  (0, 1, 1, 1, 1, 0, 0, 0, 1) 
S3_{k5 }→ S2_{k}  (0, 1, 1, 1, 0, 0, 0, 1, 1)  (0, 1, 1, 1, 1, 0, 0, 1, 1) 
S7_{k5 }→ S6_{k}  (1, 0, 0, 0, 0, 1, 1, 0, 0)  (1, 0, 0, 0, 1, 1, 1, 0, 0) 
S7_{k5 }→ S5_{k}  (1, 0, 0, 0, 0, 1, 1, 1, 0)  (1, 0, 0, 0, 1, 1, 1, 1, 0) 
S7_{k5 }→ S4_{k}  (1, 0, 0, 0, 0, 1, 1, 1, 1)  (1, 0, 0, 0, 1, 1, 1, 1, 1) 
S6_{k5 }→ S6_{k}  (1, 1, 0, 0, 0, 1, 1, 0, 0)  (1, 1, 0, 0, 1, 1, 1, 0, 0) 
S6_{k5 }→ S5_{k}  (1, 1, 0, 0, 0, 1, 1, 1, 0)  (1, 1, 0, 0, 1, 1, 1, 1, 0) 
S6_{k5 }→ S4_{k}  (1, 1, 0, 0, 0, 1, 1, 1, 1)  (1, 1, 0, 0, 1, 1, 1, 1, 1) 
S4_{k5 }→ S0_{k}  (1, 1, 1, 1, 0, 0, 0, 0, 0)  (1, 1, 1, 1, 1, 0, 0, 0, 0) 
S4_{k5 }→ S1_{k}  (1, 1, 1, 1, 0, 0, 0, 0, 1)  (1, 1, 1, 1, 1, 0, 0, 0, 1) 
S4_{k5 }→ S2_{k}  (1, 1, 1, 1, 0, 0, 0, 1, 1)  (1, 1, 1, 1, 1, 0, 0, 1, 1) 
S0_{k6 }→ S0_{k}  (0, 0, 0, 0, 0, 0, 0, 0, 0,  (0, 0, 0, 0, 1, 1, 0, 0, 0, 
 0)  0) 
S0_{k6 }→ S1_{k}  (0, 0, 0, 0, 0, 0, 0, 0, 0,  (0, 0, 0, 0, 1, 1, 0, 0, 0, 
 1)  1) 
S0_{k6 }→ S2_{k}  (0, 0, 0, 0, 0, 0, 0, 0, 1,  (0, 0, 0, 0, 1, 1, 0, 0, 1, 
 1)  1) 
S2_{k6 }→ S6_{k}  (0, 0, 1, 1, 0, 0, 1, 1, 0,  (0, 0, 1, 1, 1, 1, 1, 1, 0, 
 0)  0) 
S2_{k6 }→ S5_{k}  (0, 0, 1, 1, 0, 0, 1, 1, 1,  (0, 0, 1, 1, 1, 1, 1, 1, 1, 
 0)  0) 
S2_{k6 }→ S4_{k}  (0, 0, 1, 1, 0, 0, 1, 1, 1,  (0, 0, 1, 1, 1, 1, 1, 1, 1, 
 1)  1) 
S3_{k6 }→ S6_{k}  (0, 1, 1, 1, 0, 0, 1, 1, 0,  (0, 1, 1, 1, 1, 1, 1, 1, 0, 
 0)  0) 
S3_{k6 }→ S5_{k}  (0, 1, 1, 1, 0, 0, 1, 1, 1,  (0, 1, 1, 1, 1, 1, 1, 1, 1, 
 0)  0) 
S3_{k6 }→ S4_{k}  (0, 1, 1, 1, 0, 0, 1, 1, 1,  (0, 1, 1, 1, 1, 1, 1, 1, 1, 
 1)  1) 
S7_{k6 }→ S0_{k}  (1, 0, 0, 0, 0, 0, 0, 0, 0,  (1, 0, 0, 0, 1, 1, 0, 0, 0, 
 0)  0) 
S7_{k6 }→ S1_{k}  (1, 0, 0, 0, 0, 0, 0, 0, 0,  (1, 0, 0, 0, 1, 1, 0, 0, 0, 
 1)  1) 
S7_{k6 }→ S2_{k}  (1, 0, 0, 0, 0, 0, 0, 0, 1,  (1, 0, 0, 0, 1, 1, 0, 0, 1, 
 1)  1) 
S6_{k6 }→ S0_{k}  (1, 1, 0, 0, 0, 0, 0, 0, 0,  (1, 1, 0, 0, 1, 1, 0, 0, 0, 
 0)  0) 
S6_{k6 }→ S1_{k}  (1, 1, 0, 0, 0, 0, 0, 0, 0,  (1, 1, 0, 0, 1, 1, 0, 0, 0, 
 1)  1) 
S6_{k6 }→ S2_{k}  (1, 1, 0, 0, 0, 0, 0, 0, 1,  (1, 1, 0, 0, 1, 1, 0, 0, 1, 
 1)  1) 
S4_{k6 }→ S6_{k}  (1, 1, 1, 1, 0, 0, 1, 1, 0,  (1, 1, 1, 1, 1, 1, 1, 1, 0, 
 0)  0) 
S4_{k6 }→ S5_{k}  (1, 1, 1, 1, 0, 0, 1, 1, 1,  (1, 1, 1, 1, 1, 1, 1, 1, 1, 
 0)  0) 
S4_{k6 }→ S4_{k}  (1, 1, 1, 1, 0, 0, 1, 1, 1,  (1, 1, 1, 1, 1, 1, 1, 1, 1, 
 1)  1) 
S0_{k7 }→ S0_{k}  (0, 0, 0, 0, 0, 1, 1, 0, 0,  (0, 0, 0, 0, 1, 1, 0, 0, 0, 
 0, 0)  0, 0) 
S0_{k7 }→ S1_{k}  (0, 0, 0, 0, 0, 1, 1, 0, 0,  (0, 0, 0, 0, 1, 1, 0, 0, 0, 
 0, 1)  (0, 1) 
S0_{k7 }→ S2_{k}  (0, 0, 0, 0, 0, 1, 1, 0, 0,  (0, 0, 0, 0, 1, 1, 0, 0, 0, 
 1, 1)  1, 1) 
S2_{k7 }→ S6_{k}  (0, 0, 1, 1, 0, 0, 1, 1, 1,  (0, 0, 1, 1, 1, 0, 0, 1, 1, 
 0, 0)  0, 0) 
S2_{k7 }→ S5_{k}  (0, 0, 1, 1, 0, 0, 1, 1, 1,  (0, 0, 1, 1, 1, 0, 0, 1, 1, 
 1, 0)  1, 0) 
S2_{k7 }→ S4_{k}  (0, 0, 1, 1, 0, 0, 1, 1, 1,  (0, 0, 1, 1, 1, 0, 0, 1, 1, 
 1, 1)  1, 1) 
S3_{k7 }→ S6_{k}  (0, 1, 1, 1, 0, 0, 1, 1, 1,  (0, 1, 1, 1, 1, 0, 0, 1, 1, 
 0, 0)  0, 0) 
S3_{k7 }→ S5_{k}  (0, 1, 1, 1, 0, 0, 1, 1, 1,  (0, 1, 1, 1, 1, 0, 0, 1, 1, 
 1, 0)  1, 0) 
S3_{k7 }→ S4_{k}  (0, 1, 1, 1, 0, 0, 1, 1, 1,  (0, 1, 1, 1, 1, 0, 0, 1, 1, 
 1, 1)  1, 1) 
S7_{k7 }→ S0_{k}  (1, 0, 0, 0, 0, 1, 1, 0, 0,  (1, 0, 0, 0, 1, 1, 0, 0, 0, 
 0, 0)  0, 0) 
S7_{k7 }→ S1_{k}  (1, 0, 0, 0, 0, 1, 1, 0, 0,  (1, 0, 0, 0, 1, 1, 0, 0, 0, 
 0, 1)  0, 1) 
S7_{k7 }→ S2_{k}  (1, 0, 0, 0, 0, 1, 1, 0, 0,  (1, 0, 0, 0, 1, 1, 0, 0, 0, 
 1, 1)  1, 1) 
S6_{k7 }→ S0_{k}  (1, 1, 0, 0, 0, 1, 1, 0, 0,  (1, 1, 0, 0, 1, 1, 0, 0, 0, 
 0, 0)  0, 0) 
S6_{k7 }→ S1_{k}  (1, 1, 0, 0, 0, 1, 1, 0, 0,  (1, 1, 0, 0, 1, 1, 0, 0, 0, 
 0, 1)  0, 1) 
S6_{k7 }→ S2_{k}  (1, 1, 0, 0, 0, 1, 1, 0, 0,  (1, 1, 0, 0, 1, 1, 0, 0, 0, 
 1, 1)  1, 1) 
S4_{k7 }→ S6_{k}  (1, 1, 1, 1, 0, 0, 1, 1, 1,  (1, 1, 1, 1, 1, 0, 0, 1, 1, 
 0, 0)  0, 0) 
S4_{k7 }→ S5_{k}  (1, 1, 1, 1, 0, 0, 1, 1, 1,  (1, 1, 1, 1, 1, 0, 0, 1, 1, 
 1, 0)  1, 0) 
S4_{k7 }→ S4_{k}  (1, 1, 1, 1, 0, 0, 1, 1, 1,  (1, 1, 1, 1, 1, 0, 0, 1, 1, 
 1, 1)  1, 1) 
S0_{k8 }→ S0_{k}  (0, 0, 0, 0, 0, 1, 1, 1, 0,  (0, 0, 0, 0, 1, 1, 1, 0, 0, 
 0, 0, 0)  0, 0, 0) 
S0_{k8 }→ S1_{k}  (0, 0, 0, 0, 0, 1, 1, 1, 0,  (0, 0, 0, 0, 1, 1, 1, 0, 0, 
 0, 0, 1)  0, 0, 1) 
S0_{k8 }→ S2_{k}  (0, 0, 0, 0, 0, 1, 1, 1, 0,  (0, 0, 0, 0, 1, 1, 1, 0, 0, 
 0, 1, 1)  0, 1, 1) 
S2_{k8 }→ S6_{k}  (0, 0, 1, 1, 0, 0, 0, 1, 1,  (0, 0, 1, 1, 1, 0, 0, 0, 1, 
 1, 0, 0)  1, 0, 0) 
S2_{k8 }→ S5_{k}  (0, 0, 1, 1, 0, 0, 0, 1, 1,  (0, 0, 1, 1, 1, 0, 0, 0, 1, 
 1, 1, 0)  1, 1, 0) 
S2_{k8 }→ S4_{k}  (0, 0, 1, 1, 0, 0, 0, 1, 1,  (0, 0, 1, 1, 1, 0, 0, 0, 1, 
 1, 1, 1)  1, 1, 1) 
S3_{k8 }→ S6_{k}  (0, 1, 1, 1, 0, 0, 0, 1, 1,  (0, 1, 1, 1, 1, 0, 0, 0, 1, 
 1, 0, 0)  1, 0, 0) 
S3_{k8 }→ S5_{k}  (0, 1, 1, 1, 0, 0, 0, 1, 1,  (0, 1, 1, 1, 1, 0, 0, 0, 1, 
 1, 1, 0)  1, 1, 0) 
S3_{k8 }→ S4_{k}  (0, 1, 1, 1, 0, 0, 0, 1, 1,  (0, 1, 1, 1, 1, 0, 0, 0, 1, 
 1, 1, 1)  1, 1, 1) 
S7_{k8 }→ S0_{k}  (1, 0, 0, 0, 0, 1, 1, 1, 0,  (1, 0, 0, 0, 1, 1, 1, 0, 0, 
 0, 0, 0)  0, 0, 0) 
S7_{k8 }→ S1_{k}  (1, 0, 0, 0, 0, 1, 1, 1, 0,  (1, 0, 0, 0, 1, 1, 1, 0, 0, 
 0, 0, 1)  0, 0, 1) 
S7_{k8 }→ S2_{k}  (1, 0, 0, 0, 0, 1, 1, 1, 0,  (1, 0, 0, 0, 1, 1, 1, 0, 0, 
 0, 1, 1)  0, 1, 1) 
S6_{k8 }→ S0_{k}  (1, 1, 0, 0, 0, 1, 1, 1, 0,  (1, 1, 0, 0, 1, 1, 1, 0, 0, 
 0, 0, 0)  0, 0, 0) 
S6_{k8 }→ S1_{k}  (1, 1, 0, 0, 0, 1, 1, 1, 0,  (1, 1, 0, 0, 1, 1, 1, 0, 0, 
 0, 0, 1)  0, 0, 1) 
S6_{k8 }→ S2_{k}  (1, 1, 0, 0, 0, 1, 1, 1, 0,  (1, 1, 0, 0, 1, 1, 1, 0, 0, 
 0, 1, 1)  0, 1, 1) 
S4_{k8 }→ S6_{k}  (1, 1, 1, 1, 0, 0, 0, 1, 1,  (1, 1, 1, 1, 1, 0, 0, 0, 1, 
 1, 0, 0)  1, 0, 0) 
S4_{k8 }→ S5_{k}  (1, 1, 1, 1, 0, 0, 0, 1, 1,  (1, 1, 1, 1, 1, 0, 0, 0, 1, 
 1, 1, 0)  1, 1, 0) 
S4_{k8 }→ S4_{k}  (1, 1, 1, 1, 0, 0, 0, 1, 1,  (1, 1, 1, 1, 1, 0, 0, 0, 1, 
 1, 1, 1)  1, 1, 1) 
S0_{k9 }→ S0_{k}  (0, 0, 0, 0, 0, 1, 1, 1, 1,  (0, 0, 0, 0, 1, 1, 1, 1, 0, 
 0, 0, 0, 0)  0, 0, 0, 0) 
S0_{k9 }→ S1_{k}  (0, 0, 0, 0, 0, 1, 1, 1, 1,  (0, 0, 0, 0, 1, 1, 1, 1, 0, 
 0, 0, 0, 1  0, 0, 0, 1) 
S0_{k9 }→ S2_{k}  (0, 0, 0, 0, 0, 1, 1, 1, 1,  (0, 0, 0, 0, 1, 1, 1, 1, 0, 
 0, 0, 1, 1)  0, 0, 1, 1) 
S2_{k9 }→ S6_{k}  (0, 0, 1, 1, 0, 0, 0, 0, 1,  (0, 0, 1, 1, 1, 0, 0, 0, 0, 
 1, 1, 0, 0)  1, 1, 0, 0) 
S2_{k9 }→ S5_{k}  (0, 0, 1, 1, 0, 0, 0, 0, 1,  (0, 0, 1, 1, 1, 0, 0, 0, 0, 
 1, 1, 1, 0)  1, 1, 1, 0) 
S2_{k9 }→ S4_{k}  (0, 0, 1, 1, 0, 0, 0, 0, 1,  (0, 0, 1, 1, 1, 0, 0, 0, 0, 
 1, 1, 1, 1)  1, 1, 1, 1) 
S3_{k9 }→ S6_{k}  (0, 1, 1, 1, 0, 0, 0, 0, 1,  (0, 1, 1, 1, 1, 0, 0, 0, 0, 
 1, 1, 0, 0)  1, 1, 0, 0) 
S3_{k9 }→ S5_{k}  (0, 1, 1, 1, 0, 0, 0, 0, 1,  (0, 1, 1, 1, 1, 0, 0, 0, 0, 
 1, 1, 1, 0)  1, 1, 1, 0) 
S3_{k9 }→ S4_{k}  (0, 1, 1, 1, 0, 0, 0, 0, 1,  (0, 1, 1, 1, 1, 0, 0, 0, 0, 
 1, 1, 1, 1)  1, 1, 1, 1) 
S7_{k9 }→ S0_{k}  (1, 0, 0, 0, 0, 1, 1, 1, 1,  (1, 0, 0, 0, 1, 1, 1, 1, 0, 
 0, 0, 0, 0)  0, 0, 0, 0) 
S7_{k9 }→ S1_{k}  (1, 0, 0, 0, 0, 1, 1, 1, 1,  (1, 0, 0, 0, 1, 1, 1, 1, 0, 
 0, 0, 0, 1)  0, 0, 0, 1) 
S7_{k9 }→ S2_{k}  (1, 0, 0, 0, 0, 1, 1, 1, 1,  (1, 0, 0, 0, 1, 1, 1, 1, 0, 
 0, 0, 1, 1)  0, 0, 1, 1) 
S6_{k9 }→ S0_{k}  (1, 1, 0, 0, 0, 1, 1, 1, 1,  (1, 1, 0, 0, 1, 1, 1, 1, 0, 
 0, 0, 0, 0)  0, 0, 0, 0) 
S6_{k9 }→ S1_{k}  (1, 1, 0, 0, 0, 1, 1, 1, 1,  (1, 1, 0, 0, 1, 1, 1, 1, 0, 
 0, 0, 0, 1)  0, 0, 0, 1) 
S6_{k9 }→ S2_{k}  (1, 1, 0, 0, 0, 1, 1, 1, 1,  (1, 1, 0, 0, 1, 1, 1, 1, 0, 
 0, 0, 1, 1)  0, 0, 1, 1) 
S4_{k9 }→ S6_{k}  (1, 1, 1, 1, 0, 0, 0, 0, 1,  (1, 1, 1, 1, 1, 0, 0, 0, 0, 
 1, 1, 0, 0)  1, 1, 0, 0) 
S4_{k9 }→ S5_{k}  (1, 1, 1, 1, 0, 0, 0, 0, 1,  (1, 1, 1, 1, 1, 0, 0, 0, 0, 
 1, 1, 1, 0)  1, 1, 1, 0) 
S4_{k9 }→ S4_{k}  (1, 1, 1, 1, 0, 0, 0, 0, 1,  (1, 1, 1, 1, 1, 0, 0, 0, 0, 
 1, 1, 1, 1  1, 1, 1, 1) 


In this case, it is determined which of the two paths A and B has the higher probability. This decision can be made by comparing the sum of errors between ideal sample values (i.e., expected values) shown by the path A and actual sample values with the sum of errors between associated ideal sample values (i.e., expected values) shown by the path B and the actual sample values. [0129]

For example, suppose the state transition S[0130] 0 _{k−5→S6} _{k }should be estimated. As for this state transition, no matter whether the path A or the path B has been taken, the state at the time k−5 is S0 _{k−5 }and the state at the time k is S6 _{k}. In this case, it is by using the sum of squared errors between the expected values and the actual values y_{k−4}, y_{k−3}, y_{k−2}, y_{k−1 }and y_{k }of the read signal in the period between the times k−4 and k that it is determined which of the two paths A and B has the higher probability. As in the preferred embodiments described above, the sum of squared errors between the expected values shown by the path A and the actual values y_{k−4}, y_{k−3}, y_{k−2}, y_{k−1 }and y_{k }of the read signal in the period between the times k−4 and k is identified by Pa. On the other hand, the sum of squared errors between the expected values shown by the path B and the actual values y_{k−4}, y_{k−3}, y_{k−2}) y_{k−1 }and y_{k }of the read signal in the period between the times k−4 and k is identified by Pb. These sums of squared errors Pa and Pb (i.e., the error probabilities) are respectively given by the following Equations (21) and (22):

Pa=(y _{k−4}−0)^{2}+(y _{k−3} −C0)^{2}+(y_{k−2}−(C0+C1)) ^{2}+(y _{k−1}−(C0+C1+C2))^{2}+(y _{k}−(2×C1+C2))^{2} (21)

Pb=(y _{k−4} −C0)^{2}+(Y _{k−3}−(C0+C1))^{2}+(y _{k−2}−(C0+C1+C2))^{2}+(y _{k−1}−(C0+2×C1+C2))^{2}+(y _{k}−(C0+2×C1+C2))^{2} (22)

In this case, if Pa<<Pb, then the path A is estimated to be the more probable one. On the other hand, if Pa>>Pb, then the path B is estimated to be the more probable one. That is to say, even when a recorded code having a minimum polarity inversion interval of two is combined with the PR (C0, C1, C2, C1, C0) equalization technique, the reliability of the decoded result can also be evaluated by Pa−Pb. Also, the quality of the read signal can be evaluated (or the error rate can be estimated) based on the Pa−Pb distribution. [0131]

Furthermore, suppose white noise has been superposed on the transmission line. In that case, a state transition having the highest error probability should have a minimum Euclidean distance between the paths A and B. The sixteen state transition patterns shown in the following Table 9 should have the minimum Euclidean distance between their two paths:
[0132] TABLE 9 


State  Recording code  Recording code 
transition  (b_{k1}, . . . , b_{k}) of path A  (b_{k1}, . . . , b_{k}) of path B 

S0_{k5 }→ S6_{k}  (0, 0, 0, 0, 0, 1, 1, 0, 0)  (0, 0, 0, 0, 1, 1, 1, 0, 0) 
S0_{k5 }→ S5_{k}  (0, 0, 0, 0, 0, 1, 1, 1, 0)  (0, 0, 0, 0, 1, 1, 1, 1, 0) 
S0_{k5 }→ S4_{k}  (0, 0, 0, 0, 0, 1, 1, 1, 1)  (0, 0, 0, 0, 1, 1, 1, 1, 1) 
S2_{k5 }→ S0_{k}  (0, 0, 1, 1, 0, 0, 0, 0, 0)  (0, 0, 1, 1, 1, 0, 0, 0, 0) 
S2_{k5 }→ S1_{k}  (0, 0, 1, 1, 0, 0, 0, 0, 1)  (0, 0, 1, 1, 1, 0, 0, 0, 1) 
S2_{k5 }→ S2_{k}  (0, 0, 1, 1, 0, 0, 0, 1, 1)  (0, 0, 1, 1, 1, 0, 0, 1, 1) 
S3_{k5 }→ S0_{k}  (0, 1, 1, 1, 0, 0, 0, 0, 0)  (0, 1, 1, 1, 1, 0, 0, 0, 0) 
S3_{k5 }→ S1_{k}  (0, 1, 1, 1, 0, 0, 0, 0, 1)  (0, 1, 1, 1, 1, 0, 0, 0, 1) 
S3_{k5 }→ S2_{k}  (0, 1, 1, 1, 0, 0, 0, 1, 1)  (0, 1, 1, 1, 1, 0, 0, 1, 1) 
S7_{k5 }→ S6_{k}  (1, 0, 0, 0, 0, 1, 1, 0, 0)  (1, 0, 0, 0, 1, 1, 1, 0, 0) 
S7_{k5 }→ S5_{k}  (1, 0, 0, 0, 0, 1, 1, 1, 0)  (1, 0, 0, 0, 1, 1, 1, 1, 0) 
S7_{k5 }→ S4_{k}  (1, 0, 0, 0, 0, 1, 1, 1, 1)  (1, 0, 0, 0, 1, 1, 1, 1, 1) 
S6_{k5 }→ S6_{k}  (1, 1, 0, 0, 0, 1, 1, 0, 0)  (1, 1, 0, 0, 1, 1, 1, 0, 0) 
S6_{k5 }→ S5_{k}  (1, 1, 0, 0, 0, 1, 1, 1, 0)  (1, 1, 0, 0, 1, 1, 1, 1, 0) 
S6_{k5 }→ S4_{k}  (1, 1, 0, 0, 0, 1, 1, 1, 1)  (1, 1, 0, 0, 1, 1, 1, 1, 1) 
S4_{k5 }→ S0_{k}  (1, 1, 1, 1, 0, 0, 0, 0, 0)  (1, 1, 1, 1, 1, 0, 0, 0, 0) 
S4_{k5 }→ S1_{k}  (1, 1, 1, 1, 0, 0, 0, 0, 1)  (1, 1, 1, 1, 1, 0, 0, 0, 1) 
S4_{k5 }→ S2_{k}  (1, 1, 1, 1, 0, 0, 0, 1, 1)  (1, 1, 1, 1, 1, 0, 0, 1, 1) 


The reliabilities Pa−Pb of the 16 state transition patterns shown in Table 9 are given by the following Equations (23.1) through (23.18):[0133]

If (c _{k−8} , c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(0, 0, 0, 0, x, 1, 1, 0, 0), then Pa−Pb=(AA _{k−4} −BB _{k−4})+(BB _{k−3} −CC _{k−3}) +(CC _{k−2} −EE _{k−2})+(DD _{k−1} −FF _{k−1})+(DD _{k} −EE _{k}); (23.1)

If (c _{k−8} , c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(0, 0, 0, 0, x, 1, 1, 1, 0), then Pa−Pb=(AA _{k−4} −BB _{k−4})+(BB _{k−3} −CC _{k−3}) +(CC_{k−2} −EE _{k−2})+(EE _{k−1} −GG _{k−1})+(FF _{k} −GG _{k}); (23.2)

If (c_{k−8} , ck _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(0, 0, 0, 0, x, 1, 1, 1, 1), then Pa−Pb=(AA _{k−4} −BB _{k−4})+(BB _{k−3} −CC _{k−3})+(CC_{k−2} −EE _{k−2})+(EE _{k−1} −GG _{k−1})+(GG _{k} −JJ _{k}); (23.3)

If (c _{k−8} , c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} ,c _{k−1} , c _{k})=(0, 0, 1, 1, x, 0, 0, 0, 0), then Pa−Pb=(DD _{k−4} −EE _{k−4})+(DD _{k−3} −FF _{k−3})+(CC _{k−2} −EE _{k−2})+(BB _{k−1} −CC _{k−1})+(AA _{k} −BB _{k}); (23.4)

If (c _{k−8} , c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(0, 0, 1, 1, x, 0, 0, 0, 1), then Pa−Pb=(DD _{k−4} −EE _{k−4})+(DD _{k−3} −FF _{k−3})+(CC_{k−2} −EE _{k−2})+(BB _{k−1} −CC _{k−1})+(BB _{k} −HH _{k}); (23.5)

If (c _{k−8} , c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(0, 0, 1, 1, x, 0, 0, 1, 1), then Pa−Pb=(DD _{k−4} −EE _{k−4})+(DD _{k−3} −FF _{k−3})+(CC _{k−2} −EE _{k−2})+(HH _{k−1} −II _{k−1})+(CC _{k} −II _{k}); (23.6)

If (c _{k−8} , c _{k−7} , ck _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(0, 1, 1, 1, x, 0, 0, 0, 0), then Pa−Pb=(FF _{k−4} −GG _{k−4})+(EE _{k−3} −GG _{k−3})+(CC _{k−2} −EE _{k−2})+(BB _{k−1} −CC _{k−1})+(AA _{k} −BB _{k}); (23.7)

If (c _{k−8} , c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(0, 1, 1, 1, x, 0, 0, 0, 1), then Pa−Pb=(FF_{k−4} −GG _{k−4})+(EE_{k−3} −GG _{k−3})+(CC _{k−2} −EE _{k−2})+(BB _{k−1} −CC _{k−1})+(BB _{k} −HH _{k}); (23.8)

If (c _{k−8} , c _{k−7}, c_{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(0, 1, 1, 1, x, 0, 0, 1, 1), then Pa−Pb=(FF _{k−4} −GG _{k−4})+(EE _{k−3} −GG _{k−3})+(CC _{k−2} −EE _{k−2})+(HH _{k−1} −II _{k−1})+(CC _{k−1} II _{k}); (23.9)

If (c_{k−8} , c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(1, 0, 0, 0, x, 1, 1, 0, 0), then Pa−Pb=(BB _{k−4} −HH _{k−4})+(BB _{k−3} −CC _{k−3})+(CC _{k−2} −EE _{k−2})+(DD _{k−1} −FF _{k−1})+(DD _{k} −EE _{k}); (23.10)

If (c _{k−8} , c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(1, 0, 0, 0, x, 1, 1, 1, 0), then Pa−Pb=(BB _{k−4} −HH _{k−4})+(BB _{k−3} −CC _{k−3})+(CC_{k−2} −EE _{k−2})+(EE _{k−1} −GG _{k−1})+(FF _{k} −GG _{k}); (23.11)

If (c _{k−8} , c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(1, 0, 0, 0, x, 1, 1, 1, 1), then Pa−Pb=(BB _{k−4} −HH _{k−4})+(BB _{k−3} −CC _{k−3})+(CC_{k−2} −EE _{k−2})+(EE _{k−1} −GG _{k−1})+(GG _{k} −JJ _{k}); (23.12)

If (c_{k−8} , c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k}) =(1, 1, 0, 0, x, 1, 1, 0, 0), then Pa−Pb=(CC _{k−4} −II _{k−4})+(HH _{k−3} −II _{k−3})+(CC _{k−2} −EE _{k−2})+(DD _{k−1} −FF _{k−})+(DD _{k} −EE _{k}); (23.13)

If (c _{k−8 } , c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(1, 1, 0, 0, x, 1, 1, 1, 0), then Pa−Pb=(CC _{k−4} −II _{k−4})+(HH _{k−3} −II _{k−3})+(CC _{k−2} −EE _{k−2})+(EE _{k−1} −GG _{k−1})+(FF _{k} −GG _{k}); (23.14)

If (c _{k−8} , c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(1, 1, 0, 0, x, 1, 1, 1, 1), then Pa−Pb=(CC _{k−4} −II _{k−4})+(HH _{k−3} −II _{k−3})+(CC _{k−2} −EE _{k−2})+(EE _{k−1} −GG _{k−1})+(GG _{k} −JJ _{k}); (23.15)

If (c _{k−8} , c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(1, 1, 1, 1, x, 0, 0, 0, 0), then Pa−Pb=(GG _{k−4} −JJ _{k−4})+(EE _{k−3} −GG _{k−3})+(CC_{k−2} −EE _{k−2})+(BB _{k−1} −CC _{k−1})+(AA _{k} −BB _{k}); (23.16)

If (c _{k−8} , c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(1, 1, 1, 1, x, 0, 0, 0, 1), then Pa−Pb=(GG _{k−4} −JJ _{k−4})+(EE _{k−3} −GG _{k−3})+(CC _{k−2} −EE _{k−2})+(BB _{k−1} −CC _{k−1})+(BB _{k} −HH _{k}); (23.17)

and [0134]

If (c _{k−8} , c _{k−7} , c _{k−6} , c _{k−5} , c _{k−4} , c _{k−3} , c _{k−2} , c _{k−1} , c _{k})=(0, 1, 1, 1, x, 0, 0, 1, 1), then Pa−Pb=(GG _{k−4} −JJ _{k−4})+(EE _{k−3} −GG _{k−3})+(CC _{k−2} −EE _{k−2})+(HH _{k−1} −II _{k−1})+(CC _{k} −II _{k}), (23.18)

where c[0135] _{k }is the decoded result, k is an integer and AA_{k}, BB_{k}, CC_{k}, DD_{k}, EE_{k}, FF_{k}, GG_{k}, HH_{k}, II_{k }and JJ_{k }are given by:

AA _{k}=(y _{k}−0)^{2},

BB _{k}=(y _{k} −C0)^{2},

CC _{k}=(y _{k}−(C0+C1)^{2},

DD _{k}=(y _{k}−(C1+C2))^{2},

EE _{k}=(y _{k}−(C0+C1+C2)^{2},

FF _{k}=(y _{k}−(2×C1+C2))^{2},

GG _{k}=(y _{k}−(C0+2×C1+C2))^{2},

HH _{k}=(y _{k}2×C0))^{2},

II _{k}=(y _{k}−(2×C0+C1))^{2}

and[0136]

JJ _{k}=(y _{k}−(2×C0+2×C1+C2))^{2 }

Pa−Pb −(2×C0
[0137] ^{2}+2×C1
^{2}+C2
^{2}) that satisfies Equations (23.1) through (23.18) is derived from the decoded result c
_{k }and the standard deviation σ and average Pave are obtained from its distribution. Supposing that distribution is a normal distribution, the error probability is given by the following Equation (24):
$\begin{array}{cc}P\ue8a0\left(\sigma ,\mathrm{Pave}\right)=\mathrm{erfc}\ue8a0\left(\frac{\mathrm{Pave}}{\sigma}\right)& \left(24\right)\end{array}$

Thus, the standard deviation σ and average Pave may be used to estimate the error rate of the read signal or evaluate the quality of the read signal. [0138]

As described above, even when a recorded code having a minimum polarity inversion interval of two is used in combination with the PR (C0, C1, C2, C1, C0) equalization, the quality of the read signal can also be evaluated by the difference in probability Pa−Pb between the two paths of state transitions occurring in a predetermined period. [0139]
Embodiment 2

Hereinafter, a second specific preferred embodiment of the present invention will be described. The second preferred embodiment relates to a specific method of calculating the probabilities of respective states and the reliability Pa−Pb of the decoded result where the read signal is decoded by a PRML decoding method (e.g., the PR (1, 2, 2, 1) equalization technique described above). [0140]

As described above, when the PR (1, 2, 2, 1) equalization technique is adopted, a trellis diagram such as that shown in FIG. 4 is obtained. In the preferred embodiment described above, the probabilities L[0141] _{k} ^{S0 }through L_{k} ^{S5 }of the respective states S0 through S5 at the time k are given by the following Equations (25):

L _{k} ^{S0}=min [L _{k−1} ^{S0}+(y _{k}+3)^{2} , L _{k−} ^{S5}+(y _{k}+2)^{2}]

L _{k} ^{S1}=min [L _{k−1} ^{S0}+(y _{k}+2)^{2} , L _{k−1} ^{S5}+(y _{k}+1)^{2}]

L _{k} ^{S2} =L _{k−1} ^{S1}+(y _{k}+0)^{2}

L _{k} ^{S3}=min [L _{k−1} ^{S3}+(y _{k}−3)^{2} , L _{k−1} ^{S2}+(y _{k}−2)^{2}] (25)

L _{k} ^{S4}=min [L _{k−1} ^{S3}+(y _{k}−2)^{2} , L _{k−1} ^{S2}+(y _{k}−1)^{2}]

L _{k} ^{S5} =L _{k−1} ^{S4}+(y _{k}+0)^{2}

where L[0142] _{k−1} ^{S0 }through L_{k−1} ^{S5 }are the probabilities of the respective states S0 through S5 at the previous time k−1, y_{k }is the actual sample value at the time k, and min [xxx, zzz] is an operator indicating that the smaller one of xxx and zzz should be selected.

In this preferred embodiment, however, each branch metric (e.g., (y[0143] _{k}+3)^{2}) to be added to its associated probability (e.g., L_{k−1} ^{S0}) at the previous time k−1 is always divided by two and then y_{k} ^{2}/2 is always subtracted from the sum. In the PRML decoding method, the smallest one of the probabilities L_{k} ^{S0 }through L_{k} ^{S5 }may be selected by comparing them with each other. Accordingly, if these calculation rules are applied to all of the equations for obtaining L_{k} ^{S0 }through L_{k} ^{S5}, then the decoded result will not be affected at all. Thus, the probabilities L_{k} ^{S0 }through L_{k} ^{S5 }of the respective states S0 through S5 at the time k may be given by the following Equations (26):

L _{k} ^{S0}=min [L _{k−1} ^{S0}+(y _{k}+3)^{2}/2−y _{k} ^{2}/2, L _{k−1} ^{S5}+(y _{k}+2)^{2}/2−y _{k} ^{2}/2]

L _{k} ^{S1}=min [L _{k−1} ^{S0}+(y _{k}+2)^{2}/2−y _{k} ^{2}/2, L _{k−1} ^{S5}+(y _{k}+1)^{2}/2−y _{k} ^{2}/2]

L _{k} ^{S2} =L _{k−1 } ^{S1}+(y _{k}+0)^{2}/2−y _{k} ^{2}/2 (26)

L _{k} ^{S3}=min [L _{k−1} ^{S3}+(y _{k}−3)^{2}/2−y _{k} ^{2}/2, L _{k−1} ^{S2}+(y _{k}−2)^{2}/2−y _{k} ^{2}/2]

L _{k} ^{S4}=min [L _{k−1} ^{S3}+(y _{k}−2)^{2}/2−y _{k} ^{2}/2, L _{k−1} ^{S2}+(y _{k}−1)^{2}/2−y _{k} ^{2}/2]

L _{k} ^{S5} =L _{k−1} ^{S4}+(y _{k}+0)^{2}/2−y _{k} ^{2}/2

These Equations (26) may be expanded into the following Equations (27):[0144]

L _{k} ^{S0}=min [L _{k−1} ^{S0}+3y _{k}+{fraction (9/2)}, L _{k−1} ^{S5}+2y _{k}+2]

L _{k} ^{S1}=min [L _{k−1} ^{S0}+2y _{k}+2, L _{k−1} ^{S5} +y _{k}+½]

L _{k} ^{S2} =L _{k−1} ^{S1} (27)

L _{k} ^{S3}=min [L _{k−1} ^{S3}−3y _{k}+{fraction (9/2)}, L _{k−1} ^{S2}−2y _{k}+2]

L _{k} ^{S4}=min [L _{k−1} ^{S3}−2y _{k}+2, L _{k−1} ^{S2} −y _{k}+½]

L
_{k}
^{S5}
=L
_{k−1}
^{S4}

In this case, A[0145] _{k}, B_{k}, C_{k}, D_{k}, E_{k}, F_{k }and G_{k }are defined as follows:

A _{k}=3y _{k}+{fraction (9/2)}=(y _{k} −th 4)+(y _{k} −th 5)+(y _{k} −th 6)

B _{k}=2y _{k}+2=(y _{k} −th 4)+(y _{k} −th 5)

C _{k} =y _{k}+½=(y _{k} −th 4)

E _{k} =−y _{k}+½=(th 3 −y _{k})

F _{k}=−2y _{k}+2=(th 3 −y _{k})+(th 2 −y _{k})

G _{k}=−3y _{k}+{fraction (9/2)}=(th 3 −y _{k})+(th 2 −y _{k})+(th 1 −y _{k})

where th[0146] 1={fraction (5/2)}, th2={fraction (3/2)}, th3=½, th4=−½, th5=−{fraction (3/2)} and th6=−{fraction (5/2)}.

In this manner, when the sample value y[0147] _{k }is detected at the time k, the probabilities L_{k} ^{S0 }through L_{k} ^{S5 }of the respective states S0 through S5 at the time k may be obtained by calculating A_{k }through G_{k }through simple multiplications and additions following the Equations (27), i.e., without calculating the squared errors between the ideal values and the actual sample values. Thus, the circuit configuration of the ML decoder does not have to be so complicated.

As already described for the first preferred embodiment, the quality of the read signal may be evaluated by obtaining the difference in probability Pa−Pb between two possible state transition paths (i.e., paths A and B). However, this Pa−Pb calculation may also be a relatively simple one that includes no square calculations. Hereinafter, an alternative simplified method of calculating Pa−Pb will be described specifically. [0148]

In the first preferred embodiment described above, where the PR (1, 2, 2, 1) equalization technique is adopted, the Pa−Pb values are preferably obtained for such pairs of paths A and B as having the minimum Euclidean distance between them. As for the 16 state transitions shown in Table 2, for example, the Pa−Pb values are preferably obtained for the eight state transitions that result in Pa−Pb=±10 when Pa=0 or Pb=0. [0149]

For example, the Pa−Pb value may be obtained in the following manner for the S[0150] 0 _{k4}→S4 _{k }transition, which is one of those eight state transitions resulting in Pa−Pb=±10. In this case, the path A includes state transitions of S0→S0→S1→S2→S4 and the path B includes state transitions of S0→S1→S2→S3→S4. Then, the probability Pa of the path A may be given by:

Pa=(y _{k−3}+3)^{2}/2+(y _{k−2}+2)^{2}/230 (y _{k−1}+0)^{2}/2+(y _{k}−1)^{2}/2

On the other hand, the probability Pb of the path B may be given by:[0151]

Pb=(y _{k−3}+2)^{2}/2+(y_{k−2}+0)^{2}/2+(y _{k−1}−2)^{2}/2+(y _{k}−2)^{2}/2

In this case, by using the A[0152] _{k }through G_{k }values, the Pa−Pb may be obtained by:

Pa−Pb=(A _{k−3} −B _{k−3})+B _{k−2} −F _{k−1}+(E _{k} −F _{k})

In this manner, according to this preferred embodiment, the Pa−Pb value can be calculated by using the A[0153] _{k }through G_{k }values that are obtained through simple additions and subtractions on the sample value y_{k }and the preset values th1 through th6. Thus, the Pa−Pb value can be obtained relatively easily without performing the square calculations. As a result, the ML decoder may have a simplified circuit configuration.

The Pa−Pb values may also be calculated by using the A[0154] _{k }through G_{k }values in a similar manner for the other state transitions. The Pa−Pb values of some of the other state transitions may be obtained in the following manner:

As for state transition S[0155] 0 _{k−4}→S3 _{k}:

Pa−Pb=(A _{k−3} −B _{k−3})+B _{k−2} −F _{k−1}+(F _{k} −G _{k})

As for state transition S[0156] 2 _{k−4}→S0 _{k}:

Pa−Pb=(E _{k−3} −F _{k−3})−F _{k−2} +B _{k−1}+(A _{k} −B _{k})

As for state transition S[0157] 2 _{k−4}→S1 _{k}:

Pa−Pb=(E _{k−3} −F _{k−3})−F _{k−2} +B _{k−1}+(B _{k} −C _{k})

As for state transition S[0158] 5 _{k−4}→S4 _{k}:

Pa−Pb=(B _{k−3} −C _{k−3})+B _{k−2} −F _{k−1}+(E _{k} −F _{k})

As for state transition S[0159] 5 _{k−4}→S3 _{k}:

Pa−Pb=(B _{k−3} −C _{k−3})+B _{k−2} −F _{k−1}+(F _{k} −G _{k})

As for state transition S[0160] 3 _{k−4}→S0 _{k}:

Pa−Pb=(F _{k−3} −G _{k−3})−F _{k−2} +B _{k−1}+(A _{k} −B _{k})

As for state transition S[0161] 3 _{k−4}→S1 _{k}:

Pa−Pb=(F _{k−3} −G _{k−3})−F _{k−2} +B _{k−1}+(B _{k} −C _{k})

5 [0162]
Embodiment 3

Hereinafter, a third specific preferred embodiment of the present invention will be described with reference to FIG. 7. The third preferred embodiment relates to an optical disk drive [0163] 100 for use to decode a read signal by a PRML decoding method.

In the optical disk drive [0164] 100, a read signal, which has been read out from an optical disk 8 by an optical head 50, is amplified by a preamplifier 9. The preamplified signal is subjected to AC coupling and then input to an automatic gain controller (AGC) 10. The AGC 10 controls the gain of its input signal so that the output of a waveform equalizer 11 on the next stage will have predetermined amplitude. The gaincontrolled output signal of the AGC 10 has its waveform shaped by the waveform equalizer 11. Then, the waveformshaped output signal of the waveform equalizer 11 is supplied to both a PLL circuit 12 and an A/D converter 13.

The PLL circuit [0165] 12 generates a read clock signal that is synchronized with the read signal. The PLL circuit 12 may have the same configuration as the conventional PLL circuit shown in FIG. 1 (including the phase detector 5, LPF 6 and VCO 7). In response to the read clock signal that is supplied from the PLL circuit 12, the A/D converter 13 samples the read signal. The A/D converter 13 outputs the sampled data obtained in this manner to a digital filter 14.

The digital filter [0166] 14 has a frequency characteristic that has been defined so as to match the frequency characteristic of the read/write systems with the characteristic required by a Viterbi circuit 15. In this preferred embodiment, the characteristic required by the Viterbi circuit 15 is a PR (1, 2, 2, 1) equalization characteristic. The output filtered data of the digital filter 14 is input to the Viterbi circuit 15, which decodes the data by a maximum likelihood decoding method. More specifically, the Viterbi circuit 15 decodes the PR (1, 2, 2, 1) equalized signal by the maximum likelihood decoding, thereby outputting digital data.

The Viterbi circuit [0167] 15 outputs not only the decoded digital data but also Euclidean distances that have been calculated at respective points in time (i.e., branch metrics) to a differential metric analyzer 16. The differential metric analyzer 16 estimates possible state transitions from the digital data that has been supplied from the Viterbi circuit 15. Also, the differential metric analyzer 16 derives Pa−Pb, representing the reliability of the decoded result, from the estimated state transitions and the branch metrics, thereby estimating the error rate of the decoded result.

Hereinafter, the Viterbi circuit [0168] 15 and the differential metric analyzer 16 will be described in further detail with reference to FIG. 8. FIG. 8 is a block diagram illustrating an exemplary configuration for the Viterbi circuit 15 and differential metric analyzer 16. Sample values y_{k }that have been output from the digital filter 14 are input to a branch metric calculator 17 of the Viterbi circuit 15. The branch metric calculator 17 calculates respective branch metrics corresponding to the distances between the sample values y_{k }and their associated expected values Level_{v}. Since the PR (1, 2, 2, 1) equalization technique is adopted in this preferred embodiment, the expected values Level_{v }have seven values of 0 through 6. The branch metrics A_{k}, B_{k}, C_{k}, D_{k}, E_{k}, F_{k }and G_{k }representing the respective distances between the expected values and sample values y_{k }at the time k are defined by the following Equations (28):

A _{k}=(y _{k}−0)^{2},

B _{k}=(y _{k}−1)^{2},

C _{k}=(y _{k}−2)^{2},

D _{k}=(y _{k}−3)^{2}, (28)

E _{k}=(y _{k}−4)^{2},

F _{k}=(y _{k}−5)^{2 }

and [0169] G _{k}=(y _{k}−6)^{2 }

The branch metrics that have been calculated in this manner are input to an adder/comparator/selector [0170] 18. The probabilities (i.e., metric values) of the respective states S0 through S5 (see FIG. 4) at a current time k are obtained from the branch metrics at the current time k and the probabilities of those states S0 through S5 at the previous time k−1. The probabilities of the respective states SO through S5 at the current time k are given by the following Equations (29):

L _{k} ^{S0}=min [L _{k−1} ^{S0} +A _{k} , L _{k−1} ^{S5} +B _{k}]

L _{k} ^{S1}=min [L _{k−1} ^{S0} +B _{k} , L _{k−1} ^{S5} +C _{k}]

L _{k} ^{S2} =L _{k−1} ^{S1} +D _{k} (29)

L _{k} ^{S3}=min [L _{k−1} ^{S3} +G _{k} , L _{k−1} ^{S2} +F _{k]}

L _{k} ^{S4}=min [L _{k−1} ^{S3} +F _{k} , L _{k−1} ^{S2} +E _{k}]

L
_{k}
^{S5}
=L
_{k−1}
^{S4}
+D
_{k}

where min [xxx, zzz] is an operator indicating that the smaller one of the two values xxx and zzz should be selected. The metric values L[0171] _{k} ^{S0 }through L_{k} ^{S5 }at the time k are stored in a register 19 and will be used to calculate metric values of the respective states S0 through S5 at the next time k+1. The adder/comparator/selector 18 selects state transitions that have the minimum metric values in accordance with Equations (29). Also, based on the results of selection, the adder/comparator/selector 18 outputs control signals Sel0 through Sel3 to a path memory 20, which has a circuit configuration such as that shown in FIG. 9, in accordance with the following Inequalities (30):

If L _{k−1} ^{S0} +A _{k} ≧L _{k−1} ^{S5} +B _{k}, then Sel 0=“1”

If L _{k−1} ^{S0} +A _{k} <L _{k−1} ^{S5} +B _{k}, then Sel 0=“0”

If L _{k−1} ^{S0} +B _{k} ≧L _{k−1} ^{S5} +C _{k}, then Sel 1=“1”

If L _{k−1} ^{S0} +B _{k} <L _{k−1} ^{S5} +C _{k}, then Sel 1=“0” (30)

If L _{k−1} ^{S3} +G _{k} ≧L _{k−1} ^{S2} +F _{k}, then Sel 2=“1”

If L _{k−1} ^{S3} +G _{k} <L _{k−1} ^{S2} +F _{k}, then Sel 2=“0”

If L _{k−1} ^{S3} +F _{k} ≧L _{k−1} ^{S2} +E _{k}, then Sel 3=“1”

If L _{k−1} ^{S3} +F _{k} <L _{k−1} ^{S2} +E _{k}, then Sel 3=“0”

In response to the input control signals, the path memory [0172] 20 estimates most probable state transition paths according to the state transition rule and outputs digital decoded data c_{k }corresponding to the estimated state transition paths.

On the other hand, to evaluate the quality of the read signal, the branch metrics that have been output from the branch metric calculator [0173] 17 are input to a delay circuit 21. The output of the branch metrics to a differential metric calculator 22 is delayed for the amount of time corresponding to the time it takes for the adder/comparator/selector 18 and the path memory 20 to perform their signal processing. In the meantime, the output digital data c_{k }of the path memory 20 is input to a state transition detector 23, which detects predetermined patterns from the digital data c_{k}. Specifically, the state transition detector 23 detects data patterns corresponding to the eight state transitions given by Equations (9.1) though (9.8). When the state transition detector 23 detects the predetermined state transitions, the differential metric calculator 22 calculates the Pa−Pb values of those detected state transitions in accordance with the Equations (9.1) through (9.8).

It should be noted that the Pa−Pb values may be calculated by a method including no square calculations as described for the second preferred embodiment. In the method of the second preferred embodiment, the Pa−Pb values may be obtained without using the branch metrics that have been calculated by the branch metric calculator [0174] 17. Accordingly, in that case, the sample values y_{k }that have been output from the digital filter 14 may be directly input to the differential metric calculator 22 by way of the delay circuit 21 only. The differential metric calculator 22 may obtain the Pa−Pb values from the sample values y_{k }by the method described for the second preferred embodiment.

The Pa−Pb values that have been calculated in this manner for the predetermined state transitions detected are input to an average/standard deviation calculator [0175] 24. The average/standard deviation calculator 24 obtains and outputs the average Pave_{10 }and the standard deviation σ_{10 }of the distribution of the input Pa−Pb values. It should be noted that the average Pave_{10 }and the standard deviation σ_{10 }to be output in this case are obtained for predetermined state transitions, each having two possible paths with a minimum Euclidean distance between them (i.e., having relatively high error probabilities). According to Equation (11), the error rate of the read signal can be estimated by using the average Pave_{10 }and the standard deviation σ_{10}. That is to say, the standard deviation and the average obtained by the average/standard deviation calculator 24 may be used as indices that indicate the quality of the read signal and that are correlated with the error rate. It should be noted that the error rate may also be obtained with the average Pave_{10 }supposed to be zero because the average is expected to be approximately equal to zero.

The optical disk drive [0176] 100 according to the preferred embodiment described above has a configuration such as that shown in FIG. 7. Alternatively, the optical disk drive 100 may further include another waveform equalizer 28 having such an equalization characteristic as to allow the PLL circuit 12 to generate a clock signal more appropriately as shown in FIG. 10. Just like the optical disk drive 100 shown in FIG. 7, the optical disk drive 100 shown in FIG. 10 can also obtain the standard deviation and the average and can evaluate the quality of the read signal by using them. In addition, by separately providing two waveform equalizers for shaping the waveform in such a manner as to get the clock signal generated more appropriately and to get the read signal adapted to the PRML decoding method more suitably, respectively, a preferred read clock signal can be generated and the read signal can be decoded by the PRML decoding method more accurately. An optical disk drive like this, including two or more waveform equalizers, is disclosed in U.S. patent application Ser. No. 09/996,843, which was filed by the applicant of the present application and which is hereby incorporated by reference.

As another alternative, a read clock signal may also be generated based on the output of the A/D converter [0177] 13 (i.e., digital signal) as shown in FIG. 11. Even so, just like the optical disk drive 100 shown in FIG. 7, the optical disk drive 100 shown in FIG. 11 can also obtain the standard deviation and the average and can also evaluate the quality of the read signal by using them.

In the preferred embodiments described above, the quality of the read signal is evaluated by using the standard deviation σ and average Pave of the Pa−Pb distribution, which are output from the differential metric analyzer [0178] 16, as respective indices. Optionally, a control operation may also be carried out by using these indices (i.e., the standard deviation σ and average Pave) to improve the quality of the read signal. For example, the frequency characteristic of the waveform equalizer 11 may be modified by the frequency characteristic controller 29 shown in FIG. 12 so that the average or the standard deviation output from the differential metric analyzer 16 becomes zero or minimized. Then, the quality of the read signal can also be improved. Furthermore, as for an optical disk drive that can write information on the storage medium, recording parameters can be optimized by controlling the recording power or the degree of recording compensation (e.g., recording pulse width) so that the average or standard deviation, output from the differential metric analyzer 16, becomes zero or minimized.
Embodiment 4

Hereinafter, an optical disk drive according to a fourth specific preferred embodiment of the present invention will be described with reference to FIG. 13. [0179]

In this preferred embodiment, the differential metric analyzer [0180] 160 outputs the PRML error index MLSA (M=σ/2·d_{min} ^{2}) as defined by Equation (14). It should be noted that the PRML error index MLSA is obtained by dividing the standard deviation (or root mean square) σ of the most probable state transition path from the read signal by the Euclidean distance between the most probable and the second most probable state transition paths. The PRML error index MLSA is an index that can be used to evaluate the quality of the read signal appropriately when the PRML decoding technique is adopted.

As shown in FIG. 13, the error index MLSA that has been output from the differential metric analyzer [0181] 160 is supplied to a frequency characteristic controller 290. The frequency characteristic controller 290 optimizes the characteristics of the waveform equalizer 11 (e.g., the boost level and the boost center frequency thereof) so as to minimize the error index MLSA. For example, the frequency characteristic controller 290 may change the boost level slightly and then compare the PRML error index MLSA resulting from the original boost level with the PRML error index MLSA resulting from the slightly changed boost level. Based on the result of the comparison, the frequency characteristic controller 290 may select one of the two boost levels that has resulted in the smaller MLSA. By performing such an operation repeatedly, the frequency characteristic controller 290 can optimize the characteristics of the waveform equalizer 11 and converge the PRML error index MLSA to a minimum value.

As another alternative, the PRML error index MLSA that has been generated by the differential metric analyzer [0182] 160 may also be supplied to a focus offset searcher 291 as shown in FIG. 14. In reading a signal from the optical disk 8, the optical disk drive 100 performs a focus servo control so that the light beam emitted from the optical head 50 can always scan the information recording plane of the optical disk 8. This focus servo control is carried out by subjecting the focus actuator (not shown) of the optical head 50 to a feedback control so that the focus error signal that has been detected by a servo amplifier 91 is equalized with a predetermined target value X0 by way of a subtractor 92. In this case, the focus offset searcher 291 may output a value corresponding to the smallest PRML error index MLSA as the predetermined target value X0 to the subtractor 92. Then, the focus servo control may be carried out in such a manner as to minimize the PRML error index MLSA (i.e., to minimize the error rate). It should be noted that such a target value X0 may be searched for by detecting the PRML error index MLSA corresponding to a slightly changed target value X0 and comparing the MLSA value detected with the original MLSA value.

In this preferred embodiment, the focus target value is optimized by using the PRML error index MLSA. Alternatively, the PRML error index MLSA may also be used to optimize any other servo target value. For example, the PRML error index MLSA may also be used for tracking servo control, disk tilt control, lens spherical aberration correction and so on. [0183]

Furthermore, the present invention is also applicable to an optical disk drive including two optical heads [0184] 50 and 51 for reading a signal from the optical disk 8 and writing a signal on the optical disk 8, respectively, as shown in FIG. 15. In that case, the recording power may be controlled by reference to the PRML error index MLSA that is output from the differential metric analyzer 160. A signal to be written on the optical disk 8 is generated by a write signal generator 103 and then supplied to the signal writing optical head 51 by way of a modulator 102. The modulator 102 multiplies the write signal by an appropriate recording power P and then supplies the product to the optical head 51. In this case, the PRML error index MLSA that has been generated by the differential metric analyzer 160 may be supplied to a recording power controller 292. Then, the recording power controller 292 may determine the recording power P in such a manner that the PRML error index MLSA is minimized.

The optical disk drive [0185] 100 shown in FIG. 15 gets the read and write operations performed by the two different heads 50 and 51. Alternatively, a single head may be switched to perform the read or write operation selectively. Also, in the preferred embodiment described above, the recording power is controlled by using the PRML error index MLSA. Optionally, the width or the phase of write pulses may also be controlled by reference to the PRML error index MLSA.

Various preferred embodiments of the present invention described above provide a method for evaluating the quality of a read signal that has been decoded by a maximum likelihood decoding method, in which a most probable state transition path is selected from a number n of state transition paths that represent n probable transitions from a first state at a time k−j into a second state at a time k. In this method, supposing the probabilities of state transition of the most and the second most probable state transition paths in a predetermined period j between the times k−j and k (e.g., a sum of Euclidean distances in the predetermined period j) are represented by Pa and Pb, respectively, the reliability of the decoded result obtained in the period j is evaluated by Pa−Pb. Also, by measuring the Pa−Pb values a number of times and by obtaining the variance of the Pa−Pb distribution, error indices, which are correlated with the error rate of the digital decoded result obtained by the maximum likelihood decoding method, can be obtained as indices to the quality of the read signal. [0186]

While the present invention has been described with respect to preferred embodiments thereof, it will be apparent to those skilled in the art that the disclosed invention may be modified in numerous ways and may assume many embodiments other than those specifically described above. Accordingly, it is intended by the appended claims to cover all modifications of the invention that fall within the true spirit and scope of the invention. [0187]