5. Nonlinear Dynamic Models (Local Model Networks with OBF and FIR)
Dynamic Representations
Disadvantages of Traditional Nonlinear Dynamic Models
- One-step-ahead prediction error (series-parallel configuration) is optimized (usually).
- Tuning in simulation (parallel configuration) possible → Complex nonlinear problem.
- Simulation can even become unstable → Unreliable.
- Online adaptation thus is very risky.
NARX
- Problems with stability due to external output feedback.
- Very sensitive w.r.t. correct order and dead time.
- Interpolation → Strange effects.
- Optimization of equation error (in series-parallel). Drawbacks:
- emphasis on high frequencies
- one common denominator A(z) → identical dynamics for all inputs
NFIR
- No problems with stability because no feedback.
- Optimization of simulation error → All NARX difficulties are gone.
- New Drawback: Huge orders/dimensions necessary.
NOBF
- No problems with stability because only local feedback.
- Optimization of simulation error → All NARX difficulties are gone.
- New Drawback: Selection of filter pole(s)?
What's Bad With NARX? Interpolation of 2 ARX Models
What's Bad With NARX? Extrapolation
- Extrapolation behavior in the y(k‒i)-axes determines dynamics
- slope ≈ 0 → static
- |slope| < 1 → stable dynamics
- |slope| > 1 → unstable dynamics - Big advantage for all linearly extrapolating model architectures like local linear ...
- Extremely dangerous for polynomials.
- Most other model architectures extrapolate constantly.
What's Bad With NARX? MISO Models
- Each input → output: Individual dynamics.
- One common denominator A(z).
- Multivariate ARX model requires high order = sum of individual orders.
- Approximate zero/pole cancellations.
- Static inputs: Bi (z) needs to cancel A(z).
- Popular solution in linear case → Subspace identification in state space(FIR-based approach)
But Even With NOE? MISO Models
- OE: Linear case, each input:
- individual transfer function
- individual feedbacks
- individual states
- no. of state variables = no. of inputs x order
- NOE: Nonlinear case, each input:
- one feedback independent of inputs
- one set of states
- no. of state variables = order
Regularized FIR Estimation – The Linear Case
Recent Progress for the FIR Identification Case
- Usage of regularization for the estimation of linear FIR models (proposed by Pillonetto et. al. 2011)
Modified Loss Function for FIR Estimation
- Regularized least squares problem
- Choice of the regularization matrix
- “Tuned correlated” (TC) kernel, penalizes an exponentially decaying version of the first order derivative of the system
- Interpretation as RKHS or Gaussian Process Model
- Nonlinear optimization of the hyperparameters
Regularized FIR
- Loss function
- P specifies the parameter covariances
- P should decrease exponentially with α (α is the pole)
- P also should decrease exponentially with distances k-l.
This is the TC kernel:
3 Priors ~ Gaussian Process 3 Realizations each, order m = 20
with:
a) white, bk independent, non-decreasingb) white, bk independent, exp. decreasing
c) correlated, bk dependent, exp. decreasing
Example: Identification of Damped Second Order System
Advantages of Regularized FIR (Blue Curve)
- High variance of FIR is suppressed
- Compared to ARX the stability of the impulse response is a priori guaranteed
Extension to the Nonlinear Case – Concept
Idea
- Use regularized FIR models as local Models
- Validity functions z can depend on delayed inputs and outputs either
Advantages
- Complete model trained for simulation error
- Stability of the identified model is guaranteed
- Model order estimation is not required
Disadvantage
- Hyperparameter search for every local model
Further aspects
- Consideration of the LMN offset in the models
- Interpretation of model error as a composition of nonlinearity and noise error
Example: Saturation Type Wiener System
Estimated Results with LOLIMOT and Different Local Model Types
Next Chapter: 6. Classification Back to Overview