1 code implementation • 24 May 2024 • Jerome Sieber, Carmen Amo Alonso, Alexandre Didier, Melanie N. Zeilinger, Antonio Orvieto
While connections between these approaches exist, such models are commonly developed in isolation and there is a lack of theoretical understanding of the shared principles underpinning these architectures and their subtle differences, greatly influencing performance and scalability.
no code implementations • 24 May 2024 • Emily Cheng, Marco Baroni, Carmen Amo Alonso
The increasing prevalence of Large Language Models (LMs) in critical applications highlights the need for controlled language generation strategies that are not only computationally efficient but that also enjoy performance guarantees.
1 code implementation • 25 Mar 2024 • Carmen Amo Alonso, Jerome Sieber, Melanie N. Zeilinger
In recent years, there has been a growing interest in integrating linear state-space models (SSM) in deep neural network architectures of foundation models.
1 code implementation • 20 Mar 2023 • Jing Shuang Li, Carmen Amo Alonso
We also present an algorithm to determine the communication structure for a given system that will preserve performance while minimizing computational complexity.
2 code implementations • 22 Dec 2021 • Carmen Amo Alonso, Fengjun Yang, Nikolai Matni
By imposing locality constraints on the system response, we show that the amount of data needed for our synthesis problem is independent of the size of the global system.
1 code implementation • 3 Oct 2020 • Jing Shuang Li, Carmen Amo Alonso, John C. Doyle
The System Level Synthesis (SLS) approach facilitates distributed control of large cyberphysical networks in an easy-to-understand, computationally scalable way.