Shifeng XIE bio photo

About Me

Here is Shifeng XIE.
I am a Master’s engineering student at Telecom Paris and the Polytechnic Institute of Paris. My research interests include large language models (LLMs), time‑series foundation models, graph neural networks and image reconstruction. I am familiar with network architectures such as Transformers, Channel‑wise attention mechanisms and UNet, and with training paradigms including contrastive, masked and joint learning.

I am planning to apply for PhD positions in 2026 and welcome opportunities worldwide!
If you are interested in my work, please feel free to reach out: shifeng.xie@telecom‑paris.fr.


Research Experiences

Time Series Foundation Models

Huawei Paris Noah’s Ark Lab, France
February 2025 – August 2025

  • Designed classification time‑series foundation models and demonstrated that high‑performance pretraining is possible using only synthetic data.
  • First author of “CauKer: Classification Time Series Foundation Models Can Be Pretrained on Synthetic Data Only,” accepted by ICML 2025 Workshop on Foundation Models for Structured Data (Best Time Series Paper).
  • Collaborated with Ievgen Redko and researchers from Huawei Paris.

In‑Context Learning and Mixture of Experts

Stellantis, France
July 2024 – January 2025

  • Investigated in‑context learning and gradient descent in transformers and mixture‑of‑experts (MoE) models.
  • First author of “The Initialization Determines Whether In‑Context Learning Is Gradient Descent,” published in Transactions on Machine Learning Research (TMLR), 2025.
  • Worked closely with Rui Yuan, Simone Rossi and Thomas Hannagan.

Graph Neural Networks and Graph Representation Learning

Télécom Paris (Polytechnic Institute of Paris), France
December 2023 – September 2024

  • Conducted research on variational graph contrastive learning and subgraph Gaussian embedding for self‑supervised graph representation.
  • First author of “Variational Graph Contrastive Learning,” accepted by NeurIPS 2024 Workshop on Self‑Supervised Learning – Theory and Practice.
  • First author of “Subgraph Gaussian Embedding Contrast for Self‑Supervised Graph Representation Learning,” accepted by ECML‑PKDD 2025.
  • Supervised by Jhony H. Giraldo.

Image Processing and High‑Dynamic‑Range Reconstruction

Xidian University, China
February 2023 – September 2023

  • Developed FTUnet for single HDR image reconstruction.
  • First author of “FTUnet: Feature Transferred U‑Net for Single HDR Image Reconstruction,” accepted by ACM Multimedia Asia (MMA) 2023, oral presentation.
  • Advised by Liu Yi.

Data Twin and Intelligent Healthcare

Xidian University, China
April 2021 – September 2021

  • Researched the feasibility of intelligent healthcare based on digital twin and data mining.
  • First author of “Feasibility Study of Intelligent Healthcare Based on Digital Twin and Data Mining,” accepted by CISAI 2021.

Education

Master in Engineering – Signal Processing for Artificial Intelligence

Télécom Paris & Polytechnic Institute of Paris, France
August 2023 – Present

  • Current grade: 15.3 / 20.

Summer Exchange Program – Machine Learning

McGill University, Canada
July 2021 – August 2021

  • Achieved grade: A.

Bachelor of Engineering – Electronic Information Engineering

Xidian University, China
August 2019 – May 2023

  • GPA: 3.8 / 4.0.
  • Rank: 4 / 97.

Selected Projects

  • Pretraining tiny vision & language models (≈ 2 billion parameters): Trained mixture‑of‑experts models using 8 A100 80 GB GPUs on the C4 and ImageNet datasets.
  • Fine‑tuning MoE language models with permutation symmetries and LoRA (patented): Applied permutation symmetries and Low‑Rank Adaptation to MoE models (e.g., Mistral, DeepSeek, Qwen) to improve efficiency.
  • Unsupervised face recognition with PCA and ICA: Implemented PCA and ICA to extract and modify facial features, enhancing recognition accuracy.
  • Self‑supervised learning for medical image classification: Developed contrastive learning methods on the MedMNIST database to learn representations without annotations.
  • Neural network parameter diffusion (patented): Compressed experts from MoE models into a latent space via autoencoders and trained latent diffusion models to generate new experts.

Honors and Awards

  • Best Time Series Paper, ICML 2025 Workshop on Foundation Models for Structured Data – for “CauKer: Classification Time Series Foundation Models Can Be Pretrained on Synthetic Data Only”.
  • Champion, Huawei France Tech Arena “Light Chaser” Competition (2024).
  • National Scholarship (China, 2021) – highest national‑level scholarship with a 0.2 % acceptance rate.
  • National Second Prize, Chinese Student Academic Physics Competition (2021).
  • Northwest Division Champion & First Prize, Shaanxi Province, Chinese Student Academic Physics Competition – best result ever at Xidian University.
  • Honorable Mention, International College Students Mathematical Contest in Modeling (2021).
  • First Provincial Prize, Mathematical Contest in Modeling for University Students (2021).

Skills & Service

  • Languages: Chinese (native), English (C1), French (B2).
  • Programming: Python (PyTorch, TensorFlow, JAX, SciPy, Pandas), JavaScript, Java, C++ and C.
  • Hardware: Arduino, STM32, SolidWorks, VHDL, ARM and RISC.
  • Professional service: Reviewer for NeurIPS 2024 Workshop on Compression, COLM 2025, and NeurIPS 2025.

News and Updates

  • August 2025: “CauKer: Classification Time Series Foundation Models Can Be Pretrained on Synthetic Data Only” won the Best Time Series Paper award at the ICML 2025 Workshop on Foundation Models for Structured Data.
  • July 2025: Published “The Initialization Determines Whether In‑Context Learning Is Gradient Descent” in TMLR.
  • December 2024: “Variational Graph Contrastive Learning” accepted at the NeurIPS 2024 Workshop on Self‑Supervised Learning – Theory and Practice.
  • May 2024: Champion in Huawei France Tech Arena “Light Chaser” Competition.
  • September 2023: Presented “FTUnet: Feature Transferred U‑Net for Single HDR Image Reconstruction” at ACM Multimedia Asia 2023 (oral).