Jiayi Yuan
 
 6100 Main St
Houston, TX 77005
jy101 [at] rice.edu
About me
I’m Jiayi Yuan ([dʒa-ˈi:], 袁加熠), a Ph.D. candidate from the Department of Computer Science at Rice University, advised by Dr. Xia “Ben” Hu. I aim to build efficient machine learning algorithms and systems (MLSys) through methods like quantization, sparsity and re-parameterization while enhancing system robustness and security. My research applications span language, vision, time series, graph, and healthcare domains. Recently, I’ve been working on:
-  Efficiency problems of long-context LLMs. [KIVI] [KVBench] [Stop Overthinking] [AutoL2S] 
-  LLM post-training: finetune, RL, and evaluation. [Give Me FP32] [The Science] [DHP] 
-  LLM Agent, LLM Routing, LLM safety. [Honeypot] [LTSM] [Taylor Unswift] [LoRATK] 
Previously, I received my bachelor’s degree in computer science from Tsinghua University, where I also studied statistics as a minor.
I lived in Beijing for 22 years and in Houston for $YEAR-2022 year(s).
I am seeking full-time research scientist/engineer positions. Please feel free to contact me regarding any opportunities!
Education & Experience
-  Internship, 2025, NVIDIA 
-  Internship, 2024, Amazon 
-  Ph.D. in Computer Science, 2022 - 2026 (expected). Rice University 
-  B.Eng. in Computer Science and Technology, 2017 - 2021. Tsinghua University 
Highlights
-  “Give me FP32” studies nondeterminism, which has become a heated topic; e.g., it was recently featured in a blog post by Thinking Machines Lab. 
-  KIVI largely inspires KV Cache quantization in Huggingface and is integrated into Transformers. Full code is available here. 
-  Rice News: Large language models could be the key to better patient-trial matching - Rice CS Ph.D. student wins AMIA Best Student Paper Award. 
-  Rice News: Rice CS’ Xia Ben Hu investigates LLMs and likely applications. 
News
-  “Give Me FP32 or Give Me Death” got accepted to NeurIPS 2025 as an Oral (77 out of 21575 submissions) — numerical precision errors have become a hot topic! Code & Talk 
-  I got three papers at NAACL, ACL, and EMNLP 2025 each, wish I got to visit Albuquerque, Vienna, and Suzhou this year 
-  One survey on efficient LLM reasoning has been accepted by TMLR! Feel free to UPVOTE 
-  Check out our recent insights and discussions on LLM evaluation 
-  Two papers accepted by EMNLP 2024 (Main + Finding). See you in Miami! 
-  Check out our recent benchmarking works on KV Cache compression, time series foundation models and LLM evaluation! 
-  KIVI and SEED-GNN got accepted by ICML 2024. See you in Vienna! 
-  Our LLM-PTM paper is selected as a best student paper at AMIA 2023 
-  One paper accepted by NeurIPS 2023 
-  Joined Microsoft Accelerating Foundation Models Research program 
- …
Publications
Please refer to publications or Google Scholar.