Section: Scientific Foundations
Subspace-based identification and detection
For reasons closely related to the vibrations monitoring applications described in module 4.2 , we have been investigating subspace-based methods, for both the identification and the monitoring of the eigenstructure of the state transition matrix of a linear dynamical state-space system :
namely the defined by :
The (canonical) parameter vector in that case is :
where is the vector whose elements are the eigenvalues , is the matrix whose columns are the 's, and is the column stacking operator.
Subspace-based methods is the generic name for linear systems identification algorithms based on either time domain measurements or output covariance matrices, in which different subspaces of Gaussian random vectors play a key role [37] . A contribution of ours, minor but extremely fruitful, has been to write the output-only covariance-driven subspace identification method under a form that involves a parameter estimating function, from which we define a residual adapted to vibration monitoring [1] . This is explained next.
Covariance-driven subspace identification.
Let and:
be the output covariance and Hankel matrices, respectively; and: . Direct computations of the 's from the equations (9 ) lead to the well known key factorizations :
where:
are the observability and controllability matrices, respectively. The observation matrix is then found in the first block-row of the observability matrix . The state-transition matrix is obtained from the shift invariance property of . The eigenstructure then results from (10 ).
Since the actual model order is generally not known, this procedure is run with increasing model orders.
Model parameter characterization.
Choosing the eigenvectors of matrix as a basis for the state space of model (9 ) yields the following representation of the observability matrix:
where , and and are as in (11 ). Whether a nominal parameter fits a given output covariance sequence is characterized by [1] :
This property can be checked as follows. From the nominal , compute using (15 ), and perform e.g. a singular value decomposition (SVD) of for extracting a matrix such that:
Matrix is not unique (two such matrices relate through a post-multiplication with an orthonormal matrix), but can be regarded as a function of . Then the characterization writes:
Residual associated with subspace identification.
Assume now that a reference and a new sample are available. For checking whether the data agree with , the idea is to compute the empirical Hankel matrix :
and to define the residual vector:
Let be the actual parameter value for the system which generated the new data sample, and be the expectation when the actual system parameter is . From (18 ), we know that has zero mean when no change occurs in , and nonzero mean if a change occurs. Thus plays the role of a residual.
It is our experience that this residual has highly interesting properties, both for damage detection [1] and localization [3] , and for flutter monitoring [8] .
Other uses of the key factorizations.
Factorization ( 3.5.1 ) is the key for a characterization of the canonical parameter vector in (11 ), and for deriving the residual. Factorization (13 ) is also the key for :
-
Proving consistency and robustness results [6] ;
-
Designing an extension of covariance-driven subspace identification algorithm adapted to the presence and fusion of non-simultaneously recorded multiple sensors setups [7] ;
-
Proving the consistency and robustness of this extension [9] ;
-
Designing various forms of input-output covariance-driven subspace identification algorithms adapted to the presence of both known inputs and unknown excitations [10] .