"transformer architecture" Papers

335 papers found • Page 5 of 7

What Makes a Good Diffusion Planner for Decision Making?

Haofei Lu, Dongqi Han, Yifei Shen et al.

ICLR 2025arXiv:2503.00535
27
citations

Why In-Context Learning Models are Good Few-Shot Learners?

Shiguang Wu, Yaqing Wang, Quanming Yao

ICLR 2025

ZETA: Leveraging $Z$-order Curves for Efficient Top-$k$ Attention

Qiuhao Zeng, Jierui Huang, Peng Lu et al.

ICLR 2025arXiv:2501.14577
5
citations

A Comparative Study of Image Restoration Networks for General Backbone Network Design

Xiangyu Chen, Zheyuan Li, Yuandong Pu et al.

ECCV 2024arXiv:2310.11881
53
citations

Active Coarse-to-Fine Segmentation of Moveable Parts from Real Images

Ruiqi Wang, Akshay Gadi Patil, Fenggen Yu et al.

ECCV 2024arXiv:2303.11530
5
citations

ALERT-Transformer: Bridging Asynchronous and Synchronous Machine Learning for Real-Time Event-based Spatio-Temporal Data

Carmen Martin-Turrero, Maxence Bouvier, Manuel Breitenstein et al.

ICML 2024oralarXiv:2402.01393
6
citations

An Incremental Unified Framework for Small Defect Inspection

Jiaqi Tang, Hao Lu, Xiaogang Xu et al.

ECCV 2024arXiv:2312.08917
22
citations

A Simple Low-bit Quantization Framework for Video Snapshot Compressive Imaging

Miao Cao, Lishun Wang, Huan Wang et al.

ECCV 2024arXiv:2407.21517
4
citations

A Tale of Tails: Model Collapse as a Change of Scaling Laws

Elvis Dohmatob, Yunzhen Feng, Pu Yang et al.

ICML 2024arXiv:2402.07043
110
citations

Attention Disturbance and Dual-Path Constraint Network for Occluded Person Re-identification

Jiaer Xia, Lei Tan, Pingyang Dai et al.

AAAI 2024paperarXiv:2303.10976
24
citations

Attention Meets Post-hoc Interpretability: A Mathematical Perspective

Gianluigi Lopardo, Frederic Precioso, Damien Garreau

ICML 2024arXiv:2402.03485
14
citations

Auctionformer: A Unified Deep Learning Algorithm for Solving Equilibrium Strategies in Auction Games

Kexin Huang, Ziqian Chen, xue wang et al.

ICML 2024

AVSegFormer: Audio-Visual Segmentation with Transformer

Shengyi Gao, Zhe Chen, Guo Chen et al.

AAAI 2024paperarXiv:2307.01146
82
citations

Bidirectional Multi-Scale Implicit Neural Representations for Image Deraining

Xiang Chen, Jinshan Pan, Jiangxin Dong

CVPR 2024arXiv:2404.01547
84
citations

Breaking through the learning plateaus of in-context learning in Transformer

Jingwen Fu, Tao Yang, Yuwang Wang et al.

ICML 2024arXiv:2309.06054
5
citations

Bridging the Gap between 2D and 3D Visual Question Answering: A Fusion Approach for 3D VQA

Wentao Mo, Yang Liu

AAAI 2024paperarXiv:2402.15933
26
citations

CarFormer: Self-Driving with Learned Object-Centric Representations

Shadi Hamdan, Fatma Guney

ECCV 2024arXiv:2407.15843
12
citations

CityGuessr: City-Level Video Geo-Localization on a Global Scale

Parth Parag Kulkarni, Gaurav Kumar Nayak, Shah Mubarak

ECCV 2024arXiv:2411.06344
9
citations

Converting Transformers to Polynomial Form for Secure Inference Over Homomorphic Encryption

Itamar Zimerman, Moran Baruch, Nir Drucker et al.

ICML 2024arXiv:2311.08610
22
citations

Correlation Matching Transformation Transformers for UHD Image Restoration

Cong Wang, Jinshan Pan, Wei Wang et al.

AAAI 2024paperarXiv:2406.00629
59
citations

DiffSurf: A Transformer-based Diffusion Model for Generating and Reconstructing 3D Surfaces in Pose

Yoshiyasu Yusuke, Leyuan Sun

ECCV 2024arXiv:2408.14860
2
citations

Distilling Morphology-Conditioned Hypernetworks for Efficient Universal Morphology Control

Zheng Xiong, Risto Vuorio, Jacob Beck et al.

ICML 2024arXiv:2402.06570
7
citations

Dolfin: Diffusion Layout Transformers without Autoencoder

Yilin Wang, Zeyuan Chen, Liangjun Zhong et al.

ECCV 2024arXiv:2310.16305
26
citations

Dynamic Memory Compression: Retrofitting LLMs for Accelerated Inference

Piotr Nawrot, Adrian Łańcucki, Marcin Chochowski et al.

ICML 2024arXiv:2403.09636
94
citations

EcoMatcher: Efficient Clustering Oriented Matcher for Detector-free Image Matching

Peiqi Chen, Lei Yu, Yi Wan et al.

ECCV 2024
4
citations

EDformer: Transformer-Based Event Denoising Across Varied Noise Levels

Bin Jiang, Bo Xiong, Bohan Qu et al.

ECCV 2024
11
citations

Effective Lymph Nodes Detection in CT Scans Using Location Debiased Query Selection and Contrastive Query Representation in Transformer

Qinji Yu, Yirui Wang, Ke Yan et al.

ECCV 2024arXiv:2404.03819
2
citations

Efficient Pre-training for Localized Instruction Generation of Procedural Videos

Anil Batra, Davide Moltisanti, Laura Sevilla-Lara et al.

ECCV 2024
1
citations

EFormer: Enhanced Transformer towards Semantic-Contour Features of Foreground for Portraits Matting

Zitao Wang, Qiguang Miao, Yue Xi et al.

CVPR 2024arXiv:2308.12831
5
citations

Event Stream-based Visual Object Tracking: A High-Resolution Benchmark Dataset and A Novel Baseline

Xiao Wang, Shiao Wang, Chuanming Tang et al.

CVPR 2024arXiv:2309.14611
86
citations

Exploring Transformer Extrapolation

Zhen Qin, Yiran Zhong, Hui Deng

AAAI 2024paperarXiv:2307.10156
12
citations

FAR: Flexible Accurate and Robust 6DoF Relative Camera Pose Estimation

Chris Rockwell, Nilesh Kulkarni, Linyi Jin et al.

CVPR 2024highlightarXiv:2403.03221
18
citations

Fast Encoding and Decoding for Implicit Video Representation

Hao Chen, Saining Xie, Ser-Nam Lim et al.

ECCV 2024arXiv:2409.19429
7
citations

Fast Registration of Photorealistic Avatars for VR Facial Animation

Chaitanya Patel, Shaojie Bai, Te-Li Wang et al.

ECCV 2024arXiv:2401.11002
1
citations

Gated Linear Attention Transformers with Hardware-Efficient Training

Songlin Yang, Bailin Wang, Yikang Shen et al.

ICML 2024arXiv:2312.06635
329
citations

GeoMFormer: A General Architecture for Geometric Molecular Representation Learning

Tianlang Chen, Shengjie Luo, Di He et al.

ICML 2024arXiv:2406.16853
9
citations

GP-NeRF: Generalized Perception NeRF for Context-Aware 3D Scene Understanding

Hao Li, Dingwen Zhang, Yalun Dai et al.

CVPR 2024highlightarXiv:2311.11863
28
citations

GPSFormer: A Global Perception and Local Structure Fitting-based Transformer for Point Cloud Understanding

Changshuo Wang, Meiqing Wu, Siew-Kei Lam et al.

ECCV 2024arXiv:2407.13519
36
citations

Graph External Attention Enhanced Transformer

Jianqing Liang, Min Chen, Jiye Liang

ICML 2024arXiv:2405.21061
9
citations

Graph Generation with $K^2$-trees

Yunhui Jang, Dongwoo Kim, Sungsoo Ahn

ICLR 2024arXiv:2305.19125
1
citations

GridFormer: Point-Grid Transformer for Surface Reconstruction

Shengtao Li, Ge Gao, Yudong Liu et al.

AAAI 2024paperarXiv:2401.02292
22
citations

GRM: Large Gaussian Reconstruction Model for Efficient 3D Reconstruction and Generation

Yinghao Xu, Zifan Shi, Wang Yifan et al.

ECCV 2024arXiv:2403.14621
264
citations

Grounding Image Matching in 3D with MASt3R

Vincent Leroy, Yohann Cabon, Jerome Revaud

ECCV 2024arXiv:2406.09756
541
citations

GS-LRM: Large Reconstruction Model for 3D Gaussian Splatting

Kai Zhang, Sai Bi, Hao Tan et al.

ECCV 2024arXiv:2404.19702
251
citations

HDformer: A Higher

Dimensional Transformer for Detecting Diabetes Utilizing Long-Range Vascular Signals - Ella Lan

AAAI 2024paperarXiv:2303.11340
3
citations

How do Transformers Perform In-Context Autoregressive Learning ?

Michael Sander, Raja Giryes, Taiji Suzuki et al.

ICML 2024

How Smooth Is Attention?

Valérie Castin, Pierre Ablin, Gabriel Peyré

ICML 2024arXiv:2312.14820
29
citations

How to Protect Copyright Data in Optimization of Large Language Models?

Timothy Chu, Zhao Song, Chiwun Yang

AAAI 2024paperarXiv:2308.12247
40
citations

How Transformers Learn Causal Structure with Gradient Descent

Eshaan Nichani, Alex Damian, Jason Lee

ICML 2024arXiv:2402.14735
102
citations

HybridGait: A Benchmark for Spatial-Temporal Cloth-Changing Gait Recognition with Hybrid Explorations

Yilan Dong, Chunlin Yu, Ruiyang Ha et al.

AAAI 2024paperarXiv:2401.00271
24
citations