About Me

Welcome! I am a PhD candidate at Université Paris Cité working on improving the efficiency of Large Language Models (LLMs) during inference. My research centres on ANN search for KV‑cache compression and LLM inference optimization. I am fortunate to be advised by Prof. Themis Palpanas.

Prior to starting my PhD, I studied Mathematics at Harbin Institute of Technology, Shenzhen and worked on pattern mining. I have a strong background in mathematics and programming (Java, C++, Python) and enjoy bridging theoretical foundations with real‑world applications.

During my MSc and subsequent research assistantships I was fortunate to collaborate with outstanding scholars. At HKUST (Guangzhou) I worked with Prof. Yuyu Luo on data selection for LLM instruction tuning. I also collaborated with Prof. Lei Chen and Dr. QiYu Liu on learned index structures. At the City University of Hong Kong, under the mentorship of Prof. Zhichao Lu, I explored multi‑objective optimisation and evolutionary computation, developing LLM‑driven automatic algorithm design (AAD). These experiences reinforced my commitment to research with real societal impact.

Education

Université Paris Cité – PhD in Computer Science

Dec 2024 – present

  • Research direction: LLM inference optimization, approximate nearest‑neighbour search for KV‑cache compression.
  • Supervisor: Prof. Themis Palpanas.
  • Currently working on KV‑cache compression and LLM inference optimization.

Harbin Institute of Technology, Shenzhen – MSc in Mathematics

Sep 2021 – Jan 2024

  • Supervisors: Prof. Guoting Chen & Prof. Wensheng Gan.
  • Research interests: pattern mining and high‑utility pattern discovery.
  • Selected courses: Functional Analysis, Tensor Analysis, Real & Complex Analysis, Stochastic Processes.
  • GPA: 3.42/4.

Dalian Maritime University – BSc in Statistics (Distinction)

Sep 2017 – Jun 2021

  • Ranking: 1st out of 59 students with GPA 92.9/100.
  • Representative courses: Complex Variable Functions, Mathematical Analysis, Numerical Analysis, Probability Theory, Time Series Analysis and C++ Programming.

Experience

Research Assistant, HKUST (Guangzhou)

Oct 2024 – Dec 2024

  • Advised by Prof. Yuyu Luo on Data selection.
  • LEAD: Iterative Data Selection for Efficient LLM Instruction Tuning, 2025. [link]

Research Assistant, City University of Hong Kong

Jan 2024 – Apr 2024

  • Advised by Prof. Zhichao Lu on LLM-driven evolutionary computation and automatic algorithm design.
  • RecRanker–Instruction Tuning LLM as Ranker for Top‑k Recommendation, ACM TOIS, 2024. [link]
  • Automated Algorithm Customisation via Evolutionary Program Search with LLMs, 2025.

Visiting Student, Hong Kong University of Science & Technology

Sep 2023 – Mar 2024

  • Advised by Prof. Qiyu Liu and Prof. Lei Chen on AI‑for‑database research such as learned indexes.
  • Why Are Learned Indexes So Effective but Sometimes Ineffective? VLDB 2025. [link]
  • Not Small Enough? SegPQ: A Learned Approach to Compress Product Quantization Codebooks, VLDB 2025.

Research Interests

  • LLM inference optimization: developing efficient transformer decoding with compressed KV caches, including vector search strategies and quantisation techniques.
  • Approximate nearest‑neighbour search for KV caches: designing ANN indexes tailored for the key–value memory structures used in generative models.

Selected Publications

For a complete list see my Google Scholar profile. Here are a few highlights:

  • Xiaotian Lin, Yanlin Qi, Yizhang Zhu, Themis Palpanas, Chengliang Chai, Nan Tang, Yuyu Luo. LEAD: Iterative Data Selection for Efficient LLM Instruction Tuning. arXiv preprint 2025. [link]
  • Qiyu Liu∗†, Siyuan Han, Yanlin Qi(co-first), Jingshu Peng, Jin Li, Longlong Lin, Lei Chen. Why Are Learned Indexes So Effective but Sometimes Ineffective? VLDB 2025. [link]
  • Qiyu Liu∗†, Yanlin Qi(co-first), Siyuan Han, Jingshu Peng, Jin Li, Lei Chen. Not Small Enough? SegPQ: A Learned Approach to Compress Product Quantization Codebooks. VLDB 2025.
  • Sichun Luo, Bowei He, Haohan Zhao, Wei Shao, Yanlin Qi, Linqi Song. RecRanker: Instruction Tuning Large Language Model as Ranker for Top‑k Recommendation. ACM Transactions on Information Systems, 2023. [link]
  • Yanlin Qi, Guoting Chen, Wensheng Gan. Mining periodic trends via closed high‑utility patterns. Expert Systems with Applications, 2023. [link]
  • Yanlin Qi, Guoting Chen. Mining Valuable Fuzzy Patterns via the RFM Model. ICDMW 2022. [link]
  • Xiaojie Zhang, Yanlin Qi, Guoting Chen, Wensheng Gan. Fuzzy‑driven periodic frequent pattern mining. Information Sciences, 2022. [link]
  • Yanlin Qi, Guoting Chen, Wensheng Gan. F‑RFM‑Miner: An efficient algorithm for mining valuable fuzzy patterns via the RFM model. Applied Intelligence, 2023. [link]

Projects

Multi‑Objective Optimisation for Resource Management in Big Data Processing

Jun 2023 – Aug 2023

Transformed stage‑level multi‑objective optimisation into query‑level optimisation for resource allocation in big data workflows. Proved that the query‑level Pareto frontier can be derived from stage‑level optima and designed getQueryPOSet algorithms with reduced complexity.

Knowledge Discovery on Multi‑temporal Data

Oct 2021 – Jan 2023

Explored association rule mining on multi‑temporal databases, introducing fuzzy datasets and negative‑utility datasets. Developed specialised algorithms for periodic, closed and stable periodic patterns in sequences and graphs.

Honours & Awards

  • National Encouragement Scholarship (×3), 2018–2021.
  • First‑class Scholarship, Harbin Institute of Technology (×3), 2021–2022.
  • Excellent Master’s Thesis (top 2%) at Harbin Institute of Technology, Jan 2024.
  • Innovation & Entrepreneurship Training Programme Award, 2019–2020.
  • Excellent Student Award, Dalian Maritime University, May 2021.
  • Multiple prizes in mathematics competitions at Dalian Maritime University.