😄 About Me
Hello, I’m Christian Yang (杨智桓 / Yang Chi Wun in Cantonese). I graduated with a Bachelor of Engineering in Artificial Intelligence from Sun Yat-sen University in Fall 2025. Currently, I’m taking a gap year working as an AI Engineer at Sengine AI, a promising startup, where I focus on Retrieval-Augmented Generation (RAG) and Multi-Agent Systems.
During my undergraduate studies, I had the fortune of receiving academic guidance from Dr. Zhao Song and Dr. Yingyu Liang. My research interests include Large Language Models (efficient training & inference, reasoning, and science of LLMs), AI Security, and Deep Learning Theory. I welcome discussions, critical feedback, or novel ideas related to my work—please don’t hesitate to reach out via email at christiannyang37 [at] gmail [dot] com.
I am now actively preparing for the next phase of my career and am seeking PhD student positions or on-site research internships commencing in Fall 2026. I am open to discussing potential opportunities and collaborations.
📝 Publications & Selected Preprints
* denotes equal contribution in alphabetical order
- Timothy Chu*, Zhao Song*, and Chiwun Yang*. How to protect copyright data in optimization of large language models? AAAI 2024 Poster, arXiv preprint arXiv:2308.12247
- Yingyu Liang*, Zhenmei Shi*, Zhao Song*, and Chiwun Yang*. Towards infinite-long prefix in transformer. ICLR 2025 SCOPE Workshop Oral (Top 9% papers), arXiv preprint arXiv:2406.14036
- Zhao Song*, Jing Xiong*, and Chiwun Yang*. How sparse attention approximates exact attention? Your attention is naturally $n^C$-sparse. ICLR 2025 SLLM Workshop Poster, arXiv preprint arXiv:2404.02690
- Yang Cao*, Zhao Song*, and Chiwun Yang*. Video latent flow matching: Optimal polynomial projections for video interpolation and extrapolation. ICLR 2025 DeLTa Workshop Poster, arXiv preprint arXiv:2502.00500
- Majid Daliri*, Zhao Song*, and Chiwun Yang*. Unlocking the theory behind scaling 1-bit neural networks. CPAL 2025 Poster, arXiv preprint arXiv:2411.01663
- Yekun Ke*, Yingyu Liang*, Zhenmei Shi*, Zhao Song*, and Chiwun Yang*. Curse of attention: A kernel-based perspective for why transformers fail to generalize on time series forecasting and beyond. CPAL 2025 Poster, arXiv preprint arXiv:2412.06061
- Yichuan Deng*, Zhao Song*, Shenghao Xie*, and Chiwun Yang*. Unmasking transformers: A theoretical approach to data recovery via attention weights. arXiv preprint arXiv:2310.12462, 2023
- Jing Xiong, Jianghan Shen, Chuanyang Zheng, Zhongwei Wan, Chenyang Zhao, Chiwun Yang, Fanghua Ye, Hongxia Yang, Lingpeng Kong, Ngai Wong. ParallelComp: Parallel long-context compressor for length extrapolation. ICML 2025 Poster, arXiv preprint arXiv:2502.14317, 2025
- Jiangxuan Long*, Zhao Song*, and Chiwun Yang*. Theoretical foundation of flow-based time series generation: provable approximation, generalization, and efficiency. arXiv preprint arXiv:2503.14076, 2025
📖 Educations
- 2021.09 - 2025.06, School of Artificial Intelligence, Sun Yat-sen University.
💻 Research & Work Experience
- 2022.08 - 2024.02, Research Assistant, Shenzhen Institute of Artificial Intelligence and Robotics for Society, advised by Dr. Nan Li.
- 2022.10 - 2025.03, Research Intern, Zhao’s Lab advised by Dr. Zhao Song.
- 2024.10 - current, Research Intern, NGai’s Lab, The University of Hong Kong (HKU), advised by Jing Xiong.
- 2024.06 - current, Machine Learning Engineer, Sengine AI.`
📊 Service
- ICLR 2025, ICLR 2025 Workshop, ICML 2025, COLM 2025