top of page
IFAC 2020 World Congress Workshop on
​
"Data-Driven Modelling, Data Reconciliation and Fault Detection using Principal Component Analysis and its New Variants

This page gives you the full information on our upcoming workshop at the 21st IFAC 2020 World Congress to be held at Berlin, Germany from July 12-17, 2020.

​

What is this workshop about?

​

The focus of this workshop is on presenting the theory and tools for PCA-based approaches to EIV identification, while a brief overview of other methods will also be provided. The objective is to deliver a simple standalone technique for an automated complete identification of a process that includes order estimation, model coefficients and error variances from input-output data. Subsequently, we focus on statistical fault diagnosis using these models and residual-based approaches using EIV-Kalman filters.

Participants will be trained in developing (i) regression models for steady-state processes (ii) dynamical models  including both input-output and state-space classes for dynamic processes and (iii) fault detection using DIPCA-EIV-Kalman filter. Case studies using an in-house developed MATLAB-based GUI toolbox for model identification and fault detection using the aforementioned methods will be presented to illustrate the methods on applications relating to regression, data reconciliation and fault detection.

​

Who would be interested in this workshop?

​

Researchers and practitioners working in the areas of identification,  data reconciliation, control and monitoring.

​

Academicians who are interested in developing identification tools using multivariate data analysis and industry practitioners who are concerned with deploying near automated solutions for developing models from data and implementing fault detection techniques for processes would benefit from this workshop.

​

Why is this workshop useful?

​

Models are central to all applications of process automation including design, data reconciliation, control, optimization and process monitoring. Developing models from data, formally known as system identification, has been a powerful alternative to first-principles approaches, and in many situations the de facto choice for complex processes. Data-driven models are also advantageous in capturing effects of uncertainties, modelling random signals and estimating noise levels in process and measurements. The nearly seven decades of literature presents a rich repertoire of techniques with excellent, practically useful, software tools for identification. A majority of these techniques, however, cater to the case of error-free inputs. There exist, however, a large number of applications where the inputs are also known with errors (in addition to outputs) - identification problems in these cases are known as the errors-in-variables (EIV) identification. Techniques devised for classical identification, when applied to solve the EIV problems, are known to result in biased estimates. Furthermore, the statistical properties of input-errors need to be estimated. In this respect, EIV identification has emerged as a separate significant branch of identification in its own right (Soderstrom, 2018). EIV identification finds applications in a variety of engineering and modern applications including process industry, manufacturing, biological processes and econometrics. Techniques for EIV identification include (total) least-squares, instrumental variable, maximum likelihood and frequency-domain algorithms. In the recent times, principal component analysis (PCA)-based techniques for EIV identification, specifically, iterative and dynamic iterative PCA methods (IPCA and DIPCA), have been developed for building steady-state and dynamical models, respectively (Narasimhan and Shah, 2008; Maurya et al, 2018).

 

A key advantage of PCA-based formulation over existing methods is that, with a careful handling of measurement noise, the order of the dynamical system, model coefficients and noise variances can be estimated consistently using a layered approach. Consistency of order determination is applicable to both input-output and state-space models. Moreover, PCA-based approaches are symmetric, in the sense that models are first identified as constraints where no prior distinction of variables as input and output is required. Subsequently, the model is obtained by partitioning the variables into input and outputs and re-writing the constraints accordingly. This approach is also useful in other applications such as soft sensing, imputation of missing observations, etc. Models built using dynamic PCA carry biased estimates in general, whereas those that are developed using dynamic iterative PCA are consistent (unbiased). Furthermore, DIPCA facilitates a methodology for accurate model order determination and provides estimates of the noise covariance matrix.

 

Two of the key applications of models are in data reconciliation and in statistical process monitoring (a.k.a. fault detection). Models for data reconciliation are typically built from first-principles. Based on a published work of Narasimhan and Bhatt (2015), we demonstrate that PCA (and its iterative version) can be, however, used for both model development and data reconciliation. This is a key result since PCA serves as a single standalone tool for both model development and data reconciliation. Model-based fault detection methods rely on the key step of residual generation. DIPCA models combined with EIV-Kalman filters provide optimal residuals and result in improved fault detection as compared to standard dynamic PCA-based approaches (Mann et al 2019). This is because the standard approaches result in biased estimates  and non-unique residuals, whereas DIPCA estimates are not only accurate but also result in unique and optimal residuals when combined with the Kalman filter.  Furthermore, these methods are invariant to non-singular transformations of the data and provide consistent results, as compared to contribution charts which are currently used for diagnosis with PCA based approaches.

​

Topics:

​

​

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

Speakers:

​

  • Arun K. Tangirala, Department of Chemical Engineering, IIT Madras

    Arun K. Tangirala holds a Bachelors in Chemical Engineering and a Doctoral degree in Process Control. He is a Professor at the Department of Chemical Engineering, IIT Madras. His research is concerned with solving cutting  edge and multi-disciplinary problems of causality analysis, network reconstruction, control loop performance monitoring, multiscale identification, sparse optimization (compressive sensing)-based identification, systems biology and modern applications of data science. He is a recipient of prestigious teaching & research awards and international fellowships. In addition, he has held visiting appointments at University of Delaware, Technical University of Munich and Tsinghua University. He was the August-Wilhelm Scheer Visiting Professor at Technical University of Munich in the year 2015. Arun Tangirala was awarded the Young Faculty Recognition Award in 2010 and the Institute Research and Development Award (Junior Level) in 2014 by IIT Madras. He is the author of a comprehensive classroom text on "Principles of System Identification: Theory and Practice", published by CRC Press. He is currently an Associate Editor for the ASME Journal of Dynamics, Measurement and Control and the Editor-in-Chief of the Journal of Institution of Engineers India: Series E (Chemical and Textile Engineering). He is also an active member of ASME, IEEE and AIChE and is a faculty associate of the Robert Bosch Centre for Data Science and Artificial Intelligence at IIT Madras.
     

  • Shankar Narasimhan, Department of Chemical Engineering, IIT Madras

    Shankar Narasimhan is currently the M.S. Ananth Institute Chair Professor in the Department of Chemical Engineering at IIT Madras. He obtained his Bachelor’s degree from IIT Madras in 1982 and his MS and PhD degrees from Northwestern University, Illinois, USA in 1984 and 1987, respectively. Prior to joining IIT Madras, he was an Associate Professor in the Department of Chemical Engineering at IIT Kanpur. Shankar Narasimhan’s major research interests are in the areas of Data Mining, Process Design and Optimization, Fault Detection and Diagnosis (FDD) and Fault Tolerant Control. He is well known for his work in the area of Data Reconciliation in which he has co-authored several important papers and a book which has received critical appreciation in India and abroad. Three of his publications in AIChE Journal were included in a select list of 210 papers as having made a significant contribution to Process System Engineering. Along with Engineers India Ltd., Gurgaon he has developed a software package for applying data reconciliation and gross error detection technique to process industries. For this work, he received the Engineers India Research Award instituted at IIT Kanpur in 1994-1995. In the area of process design, he has made important and seminal contributions in the design of sensor networks, heat exchanger networks, and water distribution networks. Shankar has been a visiting professor at the Centre for Automatic Control in Nancy, France, Purdue University, Clarkson University and Texas Tech University in USA and the University of Alberta in Canada. He has also spent summer internships at Engineers India Ltd., R&D Centre in Gurgaon, Honeywell Technology Solutions Ltd., R&D Centre at Bangalore, and ABB Global Services Ltd., Bangalore as part of high-level industry-academia interactions. Along with  Raghunathan Rengaswamy of IIT Madras, he co-founded Gyan Data Pvt. Ltd. in 2011, which specializes in providing services and products in data analytics, process modeling and optimization. Shankar Narasimhan was elected as a Fellow of the Indian National Academy of Engineering in 2012. 

​

​

bottom of page