Tensor-based predictive control for extremely large-scale single conjugate adaptive optics

Baptiste Sinquin*, Michel Verhaegen

*Corresponding author for this work

Research output: Contribution to journalArticleScientificpeer-review

6 Citations (Scopus)


In this paper we propose a data-driven predictive control algorithm for large-scale single conjugate adaptive optics systems. At each time sample, the Shack–Hartmann wavefront sensor signal sampled on a spatial grid of size N × N is reshuffled into a d -dimensional tensor. Its spatial-temporal dynamics are modeled with a d -dimensional autoregressive model of temporal order p, where each tensor storing past data undergoes a multilinear transformation by factor matrices of small sizes. Equivalently, the vector form of this autoregressive model features coefficient matrices parametrized with a sum of Kronecker products between d -factor matrices. We propose an Alternating Least Squares algorithm for identifying the factor matrices from open-loop sensor data. When modeling each coefficient matrix with a sum of r terms, the computational complexity for updating the sensor prediction online reduces from OpN4 in the unstructured matrix case to Oprd N2d d 1. Most importantly, this model structure breaks away from assuming any prior spatial-temporal coupling as it is discovered from the data. The algorithm is validated on a laboratory testbed that demonstrates the ability to accurately decompose the coefficient matrices of large-scale autoregressive models with a tensor-based representation, hence achieving high data compression rates and reducing the temporal error especially for a large Greenwood per sample frequency ratio.

Original languageEnglish
Pages (from-to)1612-1626
JournalJournal of the Optical Society of America A: Optics and Image Science, and Vision
Issue number9
Publication statusPublished - 2018


Dive into the research topics of 'Tensor-based predictive control for extremely large-scale single conjugate adaptive optics'. Together they form a unique fingerprint.

Cite this