Elliptical attention

SK Nielsen, LU Abdullaev, RSY Teo… - arXiv preprint arXiv …, 2024 - arxiv.org
Pairwise dot-product self-attention is key to the success of transformers that achieve state-of-
the-art performance across a variety of applications in language and vision. This dot-product …

Unifying back-propagation and forward-forward algorithms through model predictive control

L Ren, Q Li - arXiv preprint arXiv:2409.19561, 2024 - arxiv.org
We introduce a Model Predictive Control (MPC) framework for training deep neural
networks, systematically unifying the Back-Propagation (BP) and Forward-Forward (FF) …

World Models: The Safety Perspective

Z Zeng, C Zhang, F Liu, J Sifakis… - 2024 IEEE 35th …, 2024 - ieeexplore.ieee.org
With the proliferation of the Large Language Model (LLM), the concept of World Models
(WM) has recently attracted a great deal of attention in the AI research community, especially …

[PDF][PDF] Real-Time Task Planning Improvements for LLMs: Innovations in Closed-Loop Architectures

S Desai, M Gupta, K Mehta, A Nair, P Singh - 2024 - researchgate.net
Large language models (LLMs) have made significant strides in various applications, but
optimizing their task planning capabilities remains a critical challenge. To address this, we …

Unifying Back-Propagation and Forward-Forward Algorithms through Model Predictive Control

REN Lianhai, Q Li - openreview.net
We introduce a Model Predictive Control (MPC) framework for training deep neural
networks, systematically unifying the Back-Propagation (BP) and Forward-Forward (FF) …