CV

Jiecheng Lu

jlu414@gatech.edu
+1 470-929-4016
Atlanta, GA, US

Summary

Ph.D. student in Machine Learning at Georgia Tech advised by Shihao Yang. My research focuses on developing sequence-model architectures that improve expressivity under practical compute constraints, spanning attention mechanisms, linear attention, dynamic-MLP views of sequence modeling, and time series foundation models. In the 2024-2026 period, I have published 9 papers as the main contributor at top machine learning venues, including ICLR, ICML, and NeurIPS.

Research Interests

Foundation models; efficient and expressive sequence modeling; linear attention; long-context reasoning; time series analysis; sequence modeling across NLP, vision, and scientific data.

Education

  • Ph.D. in Machine Learning
    Aug. 2023 - present
    Georgia Institute of Technology
    Advisor: Shihao Yang
    Atlanta, GA
  • M.S. in Analytics
    Aug. 2021 - Aug. 2023
    Georgia Institute of Technology
    Atlanta, GA
  • B.S. in Logistics Engineering
    Sep. 2016 - Jun. 2020
    Tianjin University
    Tianjin, China

Work Experience

  • Data Scientist Intern
    Jul. 2021 - Jul. 2022
    Tencent Medical AI Lab (JARVIS Lab)
    Shenzhen, China
    • Led development of deep learning and statistical models for medical insurance forecasting and epidemic monitoring; co-authored 2 medical time-series papers, filed 6 patents as first inventor, and delivered technical solutions for medical facilities across 4 major cities handling up to 10 million data points daily.
  • Business Analyst Intern
    Jan. 2021 - Jul. 2021
    Amazon Private Brands, Global Sourcing Team
    Shenzhen, China
    • Applied machine learning and causal inference to cost analysis and sourcing decisions; built AWS-based dashboard web apps that informed decision-making and reported savings of over USD 100,000 monthly.
  • Research Assistant
    Sep. 2020 - Jan. 2021
    Peking University, Guanghua School of Management
    Beijing, China
  • Strategic and Data Analytics Intern
    Jul. 2020 - Sep. 2020
    Country Garden Group
    Foshan, China

Publications

First-authored papers

  • HyperMLP: An Integrated Perspective for Sequence Modeling
    2026
    Jiecheng Lu, Shihao Yang
    ICML 2026
  • Free Energy Mixer
    2026
    Jiecheng Lu, Shihao Yang
    ICLR 2026
    Introduces Free Energy Mixer (FEM), which interprets attention scores as a prior and performs a free-energy readout for channel-wise selective retrieval without increasing asymptotic complexity.
  • ZeroS: Zero-Sum Linear Attention for Efficient Transformers
    2025
    Jiecheng Lu, Xu Han, Yan Sun, Viresh Pati, Yubin Kim, Siddhartha Somani, Shihao Yang
    NeurIPS 2025 Spotlight
  • Linear Transformers as VAR Models: Aligning Autoregressive Attention Mechanisms with Autoregressive Forecasting
    2025
    Jiecheng Lu, Shihao Yang
    ICML 2025
  • WAVE: Weighted Autoregressive Varying Gate for Time Series Forecasting
    2025
    Jiecheng Lu, Xu Han, Yan Sun, Shihao Yang
    ICML 2025
  • In-context Time Series Predictor
    2025
    Jiecheng Lu, Yan Sun, Shihao Yang
    ICLR 2025
  • CATS: Enhancing Multivariate Time Series Forecasting by Constructing Auxiliary Time Series as Exogenous Variables
    2024
    Jiecheng Lu, Xu Han, Yan Sun, Shihao Yang
    ICML 2024
  • ARM: Refining Multivariate Forecasting with Adaptive Temporal-Contextual Learning
    2024
    Jiecheng Lu, Xu Han, Shihao Yang
    ICLR 2024

Corresponding-Author Publications and Mentored Research Leadership

  • StretchTime: Adaptive Time Series Forecasting via Symplectic Attention
    2026
    Yubin Kim, Viresh Pati, Jevon Twitty, Vinh Pham, Shihao Yang, Jiecheng Lu*
    ICML 2026
    *corresponding author; mentored undergraduate first authors
  • CAPS: Unifying Attention, Recurrence, and Alignment in Transformer-based Time Series Forecasting
    2026
    Viresh Pati, Yubin Kim, Vinh Pham, Jevon Twitty, Shihao Yang, Jiecheng Lu*
    Under review
    *corresponding author; mentored undergraduate first authors

Talks

  • Rethinking Sequence Modeling with HyperMLP: An Integrated Architectural Perspective
    2026
    Invited Talk, Knowledge Engineering Group, Tsinghua University (hosted by Prof. Jie Tang)
    Online
    Invited talk delivered in March 2026.
  • Rethinking Sequence Modeling: LLM Scaling Laws, Expressivity-Efficiency Tradeoffs, and the Role of Architecture
    2026
    PhD Seminar, Georgia Tech Machine Learning Student Seminar
    Atlanta, GA
    Student seminar talk delivered in April 2026.
  • Rethinking Sequence Modeling: LLM Scaling Laws, Expressivity-Efficiency Tradeoffs, and the Role of Architecture
    2026
    PhD Seminar, Georgia Tech ISyE PhD Student Seminar
    Atlanta, GA
    PhD seminar talk delivered in February 2026.

Computing Grants

  • Subquadratic HyperMLP via Convolution-Based Sequence Mixing
    Mar. 2026 - Mar. 2027
    Lambda Research Grant - PI
  • Scaling Zero-Sum Linear Attention for Cross-Modal Foundation Models
    Apr. 2026 - Sep. 2026
    NVIDIA Academic Grant - awarded project
  • Advanced Transformer-based Models for Long-term Time Series Forecasting
    Sep. 2025 - Sep. 2026
    NSF ACCESS Program - awarded project

Leadership, Teaching, and Service

  • Research leadership and mentoring
    Since Spring 2025, I have organized and led weekly research discussions for a five-student Georgia Tech undergraduate Sequential AI team, mentoring students on experiment design and paper writing. As of Spring 2026, this team completed three manuscripts: one published paper (ZeroS, NeurIPS 2025 Spotlight) and two ICML 2026 submissions with undergraduate first authors and me as corresponding/final author.
  • Teaching preparation
    Independent instructor, ISyE 4031 Regression and Forecasting, Georgia Tech, Summer 2026. Participant in Georgia Tech's Tech to Teaching program and finished CETL 8717 Course Design for Higher Education.
  • Reviewing service
    NeurIPS 2024-2026, ICLR 2025-2026, ICML 2025-2026, and IEEE Internet of Things Journal.

Teaching

  • ISyE 4031 Regression and Forecasting
    2026
    Georgia Institute of Technology
    Role: Independent instructor
    Summer 2026.

Research Translation

  • CDC FluSight forecasting hub
    Fall 2024 - Present
    Submit weekly influenza hospitalization forecasts for all U.S. states to the CDC FluSight forecasting hub (entry name: Gatech-ensemble), translating time-series forecasting research into public-health decision support.

Patents

  • CN115114345B
    Feature Representation Extraction for Time-Series Data
  • CN114358186B
    Data Processing for Attention-Based Time-Series Forecasting
  • CN115130656A
    Training an Anomaly Detection Model for Time-Series Data
  • CN117010542A
    Multi-Source Multimodal Data Prediction
  • CN117012347A
    Medical Data Generation from Unordered Inputs
  • CN114676176A
    Interpretable Time-Series Forecasting