Jiecheng Lu

About

I am a Ph.D. student in Machine Learning at Georgia Tech (ISyE), advised by Prof. Shihao Yang. My research focuses on developing sequence-model architectures that improve expressivity under practical compute constraints, spanning attention mechanisms, linear attention, dynamic-MLP views of sequence modeling, and time series foundation models.


News

All news


Talks

All talks


Selected Publications

Jiecheng Lu, Shihao Yang · ICLR 2026 · 2026
Introduces Free Energy Mixer (FEM), which interprets (q,k) attention scores as a prior and performs a log-sum-exp free-energy readout to reweight values at the channel level, enabling a smooth transition from mean aggregation to selective channel-wise retrieval without increasing asymptotic complexity.
Jiecheng Lu, Xu Han, Yan Sun, Viresh Pati, Yubin Kim, Siddhartha Somani, Shihao Yang · NeurIPS 2025 (Spotlight) · 2025
Introduces Zero‑Sum Linear Attention (ZeroS), which removes the uniform zero‑order term and reweights residuals to enable stable positive/negative attention weights, allowing contrastive operations within a single layer while retaining O(N) complexity.
Jiecheng Lu, Yan Sun, Shihao Yang · ICLR 2025 (Poster) · 2025
Reformulates TSF as in‑context learning by constructing tokens of (lookback, future) task pairs, enabling Transformers to adapt predictors from context without parameter updates.

All publications