Poster "state-space models" Papers

25 papers found

Effective and Efficient Time-Varying Counterfactual Prediction with State-Space Models

Haotian Wang, Haoxuan Li, Hao Zou et al.

ICLR 2025
7
citations

Efficient Time Series Processing for Transformers and State-Space Models through Token Merging

Leon Götz, Marcel Kollovieh, Stephan Günnemann et al.

ICML 2025arXiv:2405.17951
5
citations

Evolutionary Reasoning Does Not Arise in Standard Usage of Protein Language Models

Yasha Ektefaie, Andrew Shen, Lavik Jain et al.

NEURIPS 2025

Inference of Whole Brain Electrophysiological Networks Through Multimodal Integration of Simultaneous Scalp and Intracranial EEG

Shihao Yang, Feng Liu

NEURIPS 2025

Language Models Need Inductive Biases to Count Inductively

Yingshan Chang, Yonatan Bisk

ICLR 2025arXiv:2405.20131
20
citations

Mamba Modulation: On the Length Generalization of Mamba Models

Peng Lu, Jerry Huang, QIUHAO Zeng et al.

NEURIPS 2025

MedVSR: Medical Video Super-Resolution with Cross State-Space Propagation

Xinyu Liu, Guolei Sun, Cheng Wang et al.

ICCV 2025arXiv:2509.21265

Oscillatory State-Space Models

T. Konstantin Rusch, Daniela Rus

ICLR 2025arXiv:2410.03943
23
citations

Parameter Efficient Mamba Tuning via Projector-targeted Diagonal-centric Linear Transformation

Seokil Ham, Hee-Seon Kim, Sangmin Woo et al.

CVPR 2025arXiv:2411.15224
2
citations

Physics of Language Models: Part 4.1, Architecture Design and the Magic of Canon Layers

Zeyuan Allen-Zhu

NEURIPS 2025arXiv:2512.17351
12
citations

Trained Mamba Emulates Online Gradient Descent in In-Context Linear Regression

Jiarui Jiang, Wei Huang, Miao Zhang et al.

NEURIPS 2025arXiv:2509.23779
1
citations

TREAD: Token Routing for Efficient Architecture-agnostic Diffusion Training

Felix Krause, Timy Phan, Ming Gui et al.

ICCV 2025arXiv:2501.04765
13
citations

WaLRUS: Wavelets for Long range Representation Using State Space Methods

Hossein Babaei, Mel White, Sina Alemohammad et al.

NEURIPS 2025

Benchmarking and Building Long-Context Retrieval Models with LoCo and M2-BERT

Jon Saad-Falcon, Daniel Y Fu, Simran Arora et al.

ICML 2024arXiv:2402.07440
23
citations

Can Mamba Learn How To Learn? A Comparative Study on In-Context Learning Tasks

Jong Ho Park, Jaden Park, Zheyang Xiong et al.

ICML 2024arXiv:2402.04248
107
citations

Modeling Language Tokens as Functionals of Semantic Fields

Zhengqi Pei, Anran Zhang, Shuhui Wang et al.

ICML 2024

Online Variational Sequential Monte Carlo

Alessandro Mastrototaro, Jimmy Olsson

ICML 2024arXiv:2312.12616
4
citations

Outlier-robust Kalman Filtering through Generalised Bayes

Gerardo Duran-Martin, Matias Altamirano, Alex Shestopaloff et al.

ICML 2024arXiv:2405.05646
26
citations

PAC-Bayesian Error Bound, via Rényi Divergence, for a Class of Linear Time-Invariant State-Space Models

Deividas Eringis, john leth, Zheng-Hua Tan et al.

ICML 2024

Recurrent Distance Filtering for Graph Representation Learning

Yuhui Ding, Antonio Orvieto, Bobby He et al.

ICML 2024arXiv:2312.01538
13
citations

StableSSM: Alleviating the Curse of Memory in State-space Models through Stable Reparameterization

Shida Wang, Qianxiao Li

ICML 2024arXiv:2311.14495
25
citations

State-Free Inference of State-Space Models: The *Transfer Function* Approach

Rom N. Parnichkun, Stefano Massaroli, Alessandro Moro et al.

ICML 2024

The Illusion of State in State-Space Models

William Merrill, Jackson Petty, Ashish Sabharwal

ICML 2024arXiv:2404.08819
128
citations

Transformers are SSMs: Generalized Models and Efficient Algorithms Through Structured State Space Duality

Tri Dao, Albert Gu

ICML 2024arXiv:2405.21060
1146
citations

Universality of Linear Recurrences Followed by Non-linear Projections: Finite-Width Guarantees and Benefits of Complex Eigenvalues

Antonio Orvieto, Soham De, Caglar Gulcehre et al.

ICML 2024arXiv:2307.11888
35
citations