A function (FUMI) which describes the Shannon mutual information in chromatography is derived. FUMI covers two subjects in one formula: (i) the degree of peak overlap and the noise level; (ii) the mathematical formalism of peak deconvolution and quantitation based on the one-dimensional Kalman filter for peak resolution. Computer simulation demonstrates that a sufficient amount of mutual information can be retrieved from overlapped peaks (Gaussian) through (i) observation and (ii) data processing. A practical advantage of FUMI is logical evaluation and optimization of chromatographic experiments without resort to experience: the optimal can be defined as the chromatogram which can transmit the maximal amount of mutual information in a unit time.
KeywordsInformation theory, Kalman filter, mutual information, chromatography, optimizationAnalytical systems provide us with information in various forms and ways.An important aim in analytical chemistry is to elaborate a method through which more information can be transmitted from samples of interest. Logical evaluation of an overall analytical system, therefore, should be founded on the quantitative description of the information.As an example, high performance liquid chromatography (HPLC) is taken here. We focus our discussion on quantitative analysis. We can easily guess that superior chromatographic separation of target materials will yield more information.However, the amount of useful information is not only regulated by the elution pattern: reliability of the information also depends on analytical powers of data processing because of the inevitable noise-contamination, baseline drift, etc., in the measurement process.The Shannon mutual information will serve for the quantitative evaluation of the system (see below), but should cover at least two subjects: (i) the degree of peak overlap in a noise level; (ii) mathematical formalism of a data processing.We review the fundamental concepts of information theory to clarify our reasoning. Let 1[Q] be the amount of information concerning an unknown quantity Q to be estimated. If the probability of Q is known or welldefined, I[SO] can be calculated according to the Shannon equation.' We cannot know any details of the quantity Q before measurement of Q, but the complete or undistorted transmission of the information I[Q] through analytical media is always thwarted by various error sources such as calibration, interference, drift, etc. In general, the useful information which we can actually collect from an ensemble !P of the measure- If ambiguity is absent (1[Q/ !PJ 0), then we can retrieve from the measurements !V all of the original information on the unknown quantity Q. In other words, we can establish the estimation for Q with 100% certainty. If I[Q/ VI takes the largest value (=1[Q]), then nothing can be known concerning the information source Q. Chromatography achieves a decrease in the ambiguity I[Q/ !P] by separating peaks of interest, increasing the sensitivity of detection systems, etc. Data processing is also...