Data Mining Analysis of a Complex Polymer Process

Share Embed


Descrição do Produto

Communications To improve the prediction performance, more calibration data can be gathered or an on-line adjustment with gas-chromatographic measurements can be done to correct for the prediction error between the nonlinear calibration model and the true mole fractions. Control applications will definitely benefit from such improvements.

The author would like to thank Prof. Engell, Dr. Gesthuisen and S. Krämer for their valuable support and advice and his colleagues from the Lehrstuhl für Thermische Verfahrenstechnik. The design and implementation of the support vector machines was performed with the Matlab toolbox of Steve R. Gunn [7], University of Southampton. Received: February 17, 2004 [K 3212]

References

[2] [3] [4] [5] [6] [7]

By Rolf Burghaus*, Daniel Leineweber, and Jörg Lippert

1 Problem Statement

Acknowledgement

[1]

Data Mining Analysis of a Complex Multistage Polymer Process

K. Dadhe, S. Engell, R. Gesthuisen, S. Pegel, M. Völker, Proc. of the 15th IFAC World Congr. on Automatic Control, Barcelona, Paper We-M11±2, 2002. O. Höfer, Experimentelle Analyse der Methylacetatsynthese in einer katalytischen Rektifikationskolonne (in German), Diploma Thesis, Universität Dortmund 2000. V. N. Vapnik, The Nature of Statistical Learning Theory, Springer Verlag 1999. N. Cristianini, J. Shawe-Taylor, An Introduction to Support Vector Machines, Cambridge University Press, Cambridge 2000. B. Efron, R. J. Tibshirani, An Introduction to the Bootstrap, Chapman & Hall, London 1993. A. C. Davison, D. V. Hinkley, Bootstrap Methods and their Application, Cambridge University Press, Cambridge 1997. Image, Speech and Intelligent Systems, Department of Electronics and Computer Science, University of Southampton: http://www.isis.ecs.soton.ac.uk/resources/svminfo/

This paper was also published in German in Chem. Ing. Tech. 2003, 75 (6), 753.

______________________

Especially in the highly competitive commodities market, the chemical process industries (CPI) are forced to continually improve the efficiency and quality of their processes. To stay in this market, one has to produce the quality demanded by the customer in a very cost-effective manner. Therefore, many companies have invested in new systems for process and quality data acquisition in recent years. The large amounts of data collected are, however, often not adequately used in practice, because suitable automated analysis methods are only rarely available in the plants. By the successful analysis of a complex multistage polymer process we show that data mining can be used to extract valuable information from existing process and quality data. Data mining comprises various computational methods, which are used to automatically identify interesting patterns and possible relationships in large data sets. Hypotheses are generated from the data in the form of explicit rules, which can then directly be interpreted in the application context. Thereby it is often possible to discover hidden information in the data even in cases, where classical (hypothesis-based) analysis methods normally fail. The polymer process considered here consists of the two main process stages polymerization and processing, plus a number of auxiliary facilities (see Fig. 1). The polymerization stage itself is composed of several sequentially coupled units (batch/continuous), which need not be described in further detail here. On the processing stage, we consider two processing machines (MA1, MA2), which are operated in parallel. The auxiliary units are used, e.g., for solvent recovery and additive preparation. A central process information management system (PIMS) continuously collects and archives data for about 300 process parameters from polymerization, processing, and auxiliary facilities; the sampling rate is in the order of seconds. In addition, 8 quality parameters for the final polymer product are determined in the lab; only one measurement per day is available. Starting point of our data mining analysis was the need to quickly identify the cause of the large fluctuations of product quality observed over a six-month period, here shown for one of the quality parameters (see upper part of Fig. 2). Interestingly, there was a restricted time period within the analysis horizon, where the processing machines MA1 and MA2 showed significant differences regarding the quality ± [*]

950

 2004 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim

Dr. R. Burghaus ([email protected]), Dr. D. Leineweber, Dr. J. Lippert, Bayer Technology Services GmbH, Process Technology Division, D-51368 Leverkusen, Germany. DOI: 10.1002/ceat.200403206

Chem. Eng. Technol. 2004, 27, No. 9

Communications finance and insurance industries (e.g., assessment of credit-worthiness, fraud detection). More recently, a number of interesting new areas of application specifically related to the CPI have emerged: tailored data mining technologies in combination with statistical methods and neural networks are being used successfully, e.g., at Bayer group, for process data analysis and also in catalyst and life-sciences research. The data mining approach to process data analysis is particularly interesting in the context of troubleshooting activities, because it allows to quickly identify possible causes for process upsets or quality problems from available process data. Figure 1. Schematic depiction of the polymer process and characterization of the available Our newly developed data mining-driven proprocess and quality data (number of measured parameters, sampling rates). cess analysis strategy is built around the socalled subgroup discovery method [3]. This method allows to automatically identify those subsets of data records (subgroups) showing ªinterestingº deviations from the whole set of all data records regarding some target attribute. At the same time, the relevant influence factors, which characterize the subgroups, are determined. The target attribute (dependent variable, here a quality parameter) and all attributes to be considered as potential influence factors (independent variables, here all process parameters monitored by the PIMS) must be specified in advance. In addition, continuous influence factors must be suitably discretized, because the method can handle only discrete attribute values. The discretization is done by simply splitting the continuous ranges into a (usually small) number of subintervals, e.g., for high, medium and low attribute values. The analysis of the polymer process was complicated by the fact that it was not possible to directly merge the process and quality data due to the different sampling rates involved. Hence, a data reduction had to be performed for the high resolution process data during data preprocessing (DPP). To this end, the analysis horizon was divided into Figure 2. Time profile of one of the quality parameters, classified according to suitable time intervals, and each process parameter was charthe processing machines MA1 and MA2 (upper part), daily averages of the acterized on each time interval by a number of quasi-stationquality-relevant MA parameter identified by the analysis (lower part). In the ary descriptor parameters (e.g., average, standard deviation, marked time period between day 35 and day 50, the machines MA1 and MA2 produce significantly different quality. Fourier modes, event counts). The resulting quasi-stationary process data could then be combined with the corresponding quality data. For each data mining analysis, all generated deproduced. This was another fact, which had to be explained. scriptor parameters were considered as potential influence The overall objective of the analysis was, therefore, to identify the key process parameters determining quality and the factors and one of the quality parameters was chosen as tarcorresponding cause-and-effect relationships, thus providing get attribute. All continuous input parameters were discrethe basis for an improved process control. tized using eight subintervals; the interval boundaries were determined such that the same number of data records was assigned to each of the subintervals. For the target attribute, 2 Methods Used an equidistant discretization with ten subintervals was employed. Based on the preprocessed data, the subgroup discovery Data mining has its roots in the field of databases and artificial intelligence; there has been a rapid development in method generated a large number of rules, i.e., possible relarecent years [1,2]. Classical areas of application are in martionships between process and quality parameters. An autoketing (customer relationship management) and in the matic rule stability analysis ensured that only rules, which Chem. Eng. Technol. 2004, 27, No. 9

http://www.cet-journal.de

 2004 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim

951

Communications are stable with respect to certain shifts of the discretization boundaries, were considered, thereby eliminating meaningless ªdummyº rules to a large extent. For instance, the following interesting rule was found (here slightly simplified for illustration): IF Average (MA Parameter) ªHighº AND [...] THEN Quality Parameter ªHighº This rule can be nicely visualized with the underlying data (see lower part of Fig. 2); it explains the quality differences between the processing machines MA1 and MA2 observed during a restricted time period (see upper part of Fig. 2). In principle, it would have been possible to find a rule like this by direct visual inspection of the preprocessed data, but this would have been very cumbersome due to the large number of potential influence factors. For more complex rules, which describe the combined influence of several parameters, the visual approach quickly becomes hopeless. In the course of our data mining analysis, it turned out that the subgroup discovery method alone did not lead to the desired result. Basically, this was due to the fact that there were strong correlations among the input parameters (a quite common situation in process data analysis). Hence, an extremely large number of rules was generated, most of which were redundant. In order to cope with this problem, we had to find a way of filtering out a minimal set of independent influence factors which on one hand allow a sufficiently accurate description of the target attribute, on the other hand can be plausibly interpreted and provide some means for an improved process control. To this end, the subgroup discovery method was combined with the validation through a neural network in an iterative analysis workflow as follows: 1. Subgroup discovery to identify dominant influence factors (ªbest descriptorsº); 2. Manual selection of the most plausible ªbest descriptorº in case of correlated rules; 3. Neural network training using the ªbest descriptorsº found so far as input parameters; 4. Comparison of neural network prediction and target measurement: in case of sufficient agreement stop, otherwise continue with step 1. With a moderate manual effort, this iterative analysis workflow leads to a meaningful list of relevant influence factors, which are largely uncorrelated. Furthermore, the neural network provides a data-based process model, which can later be used for quality prediction. Our complete data mining workflow for process data analysis is summarized again in Fig. 3. The DPP methods employed and the subgroup discovery method are part of our ªData Mining Expert Toolboxº (DAMINEX), an inhouse development on the basis of MATLAB (The Math Works, Inc.). For the neural network, we have used the commercial software NN-Tool [4].

952

Figure 3. Iterative data mining workflow for process data analysis, comprising the two phases data preprocessing (DPP) and analysis. The neural network generated in the analysis phase can later be used for prediction of the quality parameters.

3 Results The above data mining workflow ultimately led to only four relevant influence factors for the quality parameter shown in Fig. 2: the averages of two relative feed streams and the standard deviation of a third relative feed stream (polymerization) as well as the average of a machine parameter (processing). A neural network, which was trained with these four input parameters already, provided an excellent description of the quality parameter, see upper part of Fig. 4 (only four hidden neurons were used). It turned out that almost the entire observed variation of the quality parameter could be explained by the variation of the four identified key influence parameters (model prediction and measured data were correlated with a correlation coefficient of 0.88). In addition the neural network allowed establishing a ranking of the influence factors. The effect of the MA parameter has already been discussed in the context of Fig. 2; the plant personnel confirmed the corresponding influence as highly plausible. Similarly, the influence of the relative feed rates on quality was perfectly in line with the existing process knowledge. However, there immediately arose the question, why the

 2004 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim

http://www.cet-journal.de

Chem. Eng. Technol. 2004, 27, No. 9

Communications relative feed rates ± and hence the recipe of the polymer ± showed such large variations. In order to answer this question, one of the fluctuating relative feed rates was taken as a new target attribute, and another data mining analysis was performed according to the scheme above (see Fig. 3). Here it should be noted that the relative feed rate chosen serves as the manipulated variable for viscosity control of the polymer solution produced. Relevant influence factors for this recipe parameter were the viscosity of the polymer solution (a trivial result), and, more interestingly, the binary operation mode of the auxiliary unit ªsolvent recyclingº: under high load, a second distillation train is used in parallel to the normal train. Hence, this operation mode can be characterized by a discrete parameter with only two possible values (on/off). The relative feed rate could be well predicted by a neural network trained with just the two input parameters mentioned (see lower part of Fig. 4); the corresponding correlation coefficient between model and data was 0.71. Identifying the influence of the solvent recycling unit turned out to be crucial for finding the root cause of the quality problems. It was possible to show by lab analyses

that the solvent coming from the second parallel distillation train had a higher level of impurities than the solvent coming from the normal train. Obviously, the impurities contained in the solvent led to a ªpoisoningº of the polymerization reaction and thus, via the viscosity control, to the observed recipe and quality variations.

4 Summary and Conclusions By successfully analyzing a complex polymer process, we have shown that detailed data mining analyses can be done not only for single isolated process stages, but also for large multistage processes. Data mining allows the quick identification of hidden non-local effects, which are difficult to discover with other methods. A particular advantage of data mining methods like subgroup discovery is the fact that the generated explicit rules can be interpreted directly in the application context. A specifically tailored preprocessing strategy (quasi-stationary description on suitable time intervals) allows combining process and quality data with different time resolutions for the analysis. The combination of subgroup discovery and neural network validation within an iterative analysis workflow helps to quickly narrow down on a ªmeaningfulº set of independent rules describing the target attribute of interest. Of course, our data mining methodology can be directly applied to other processes. As demonstrated by a growing number of successful applications, data mining provides a very powerful new tool for process data analysis, especially for troubleshooting. Hence, data mining is the method of choice for the automated identification of special causes in the context of statistical process control [5]. By quickly identifying the quality-relevant key influences in the process and the corresponding causeand-effect relationships, data mining can significantly contribute to process insight. Received: December 12, 2003 [K3206]

References [1] [2] [3] Figure 4. Prediction quality for two of neural network models established as part of the analysis: description of the quality parameter as a function of the MA parameter and three polymer recipe parameters (upper part), description of the main recipe parameter as a function of the binary operation mode of the auxiliary unit ªsolvent recoveryº and the viscosity of the polymer solution (lower part). The legend shows the correlation coefficient R between model prediction and measured data.

[4] [5]

Advances in Knowledge Discovery and Data Mining (Eds: U. M. Fayyad, G. Piatetsky-Shapiro, P. Smyth, R. Uthurusamy), MIT Press, Cambridge 1996. S. Wrobel, Künstliche Intelligenz 1998, 12 (1), 6. S. Wrobel, in Principles of Data Mining and Knowledge Discovery (Eds: H. J. Komorowski, J. M. Zytkow), Lecture Notes in Computer Science 1263, Springer, Berlin 1997, 78. F. Bärmann, F. Biegler-König, Neural Networks 1992, 5(1), 139. G. Box, A. Luceno, Statistical Control by Monitoring and Feedback Adjustment, 1st ed., Wiley, New York 1997.

This paper was also published in German in Chem. Ing. Tech. 2003, 75 (7), 897.

______________________

Chem. Eng. Technol. 2004, 27, No. 9

http://www.cet-journal.de

 2004 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim

953

Lihat lebih banyak...

Comentários

Copyright © 2017 DADOSPDF Inc.