Xunyi Jiang

Xunyi Jiang 江迅一

Master of Science in CS @ 🇺🇸UCSD
BS in Data Science @ 🇨🇳SUSTech

Biography

Hi! 👋🏻 I am Xunyi Jiang, a master student of Computer Science and Engineering at UCSD. I graduated from SUSTech, being recognized as one of the top ten graduates from the School of Science at SUSTech and Oustanding Thesis.

Physicists seek order in chaos through equations, while statisticians find meaning where equations fail.

My research journey began with using statistical models to explain the development of the scientific community—science for science—under the guidance of Professor Yifang Ma, related with hierarchy and mobility.

Inspired by how models can reveal hidden patterns in science, I now harness AI to drive transformative change. I host the self-supervised seminar, which is supported by Chao Wang and joined the NUS-WING group as research assistant from Jan to May 2024, advised by Min-Yen Kan. Currently, I am working with Prof. Julian Mcauley on catastrophic forgetting in LLMs.

I love music! This is one of my passion apart from coding and math. I play both cello and piano, and I used to be a principle cellist of the SUSTech Harmonic Symphonic Orchestra. I also produce music and drawing in my free time.

Interests
  • Audio-Vision Models
  • AI for Creativity 🎹
  • Controlable AI Generation
  • Self-Supervised Learning
  • Catastrophic Forgetting in LLMs
Education
  • Master of Science in Computer Science, 2024-2026(Estimated)

    University of California, San Diego

  • BS in Data Science, 2020-2024

    Southern University of Science and Technology

  • High School, 2017-2020

    No.1 Affiliated Middle School of Central China Normal University

Skills

Technical
LLMs, PEFT, Transformers🤗, PyTorch
Triton, CUDA, C++, Git, Docker
Data visualization, JavaScript, SQL, Data Dashboard, Pandas, Numpy, R
Hobbies
Cello & Piano
Hiking
Photography

Research

 
 
 
 
 
University of California, San Diego
Research Assistant
February 2025 – Present San Diego, CA

Responsibilities include:

A comprehensive experiments for catasrophic forgetting in LLMs with task-specific fine-tuning.

  • Deployed PEFT and LoRA for efficient fine-tuning.
  • Using Bayesian methods to analyze the catasrophic forgetting.
 
 
 
 
 
National University of Singapore
Research Assistant
January 2024 – May 2024 Singapore

Responsibilities include:

CLUE-ReDial: Leveraging Large Language Models for Generating Comprehensive Dataset with UrLs and Explanation

  • Designed and implemented a multi-agent framework leveraging RAG and LLMs, achieving a 4× efficiency boost in generating interpretable recommendation explanations. Expanded the ReDial dataset to 40k high-quality utterances for fine-tuning conversational LLMs, enhancing domain-specific accuracy in movie recommendations.
  • Designed a multi-agent framework, leveraging an early-stop module in the self-refine process to generate comprehensive and interpretable recommendation explanations, boosting transparency in recommendations.
  • Applied web crawling techniques to create a movie dataset containing 25k entries, using Python, PyTorch, LaTeX, and the Singapore National High-Performance Computing (HPC) platform.
 
 
 
 
 
Research Assistant
June 2022 – August 2023 Shenzhen

Responsibilities include: Worked on the OpenAlex and ORCID databases to analyze the relationship between university rankings and the mobility of researchers.

  • Extracted education and hiring networks from research paper publications.
  • Ranked Chinese universities based on extracted database information.
  • Implemented a Multi-Variate Regression (MVR) model for ranking.
  • Applied Markov Chain Monte Carlo (MCMC) and burn-in techniques to ensure the stability of convergence.
  • Visualized researcher mobility and constructed a null model to compare the hierarchical structures of research institutions.

Accomplish­ments

Top 10 undergraduates of the School of Science
Top 10 undergraduates of the School of Science and Candidate of Top Ten Graduates
National Scholarship, 2023

Recent Posts

How to merge external knowledge into LLMs

How to merge external knowledge into LLMs

Xunyi Jiang Southern University of Science and Technology xunyijiang001@gmail.com

This report explores the evolving landscape of Conversational Recommender Systems (CRS) in the context of Large Language Models (LLMs). It delves into the integration of knowledge graphs and advanced computational techniques to enhance the capabilities of CRS. The report categorizes current CRS methods into four main classes: knowledge-based, goal-driven, fine-tuning, and agent-based approaches, providing an in-depth analysis of each. Special attention is given to how these systems can optimize conversations and recommendations by harnessing the power of LLMs and external knowledge sources. The innovative use of agent-based methods and the challenges of fine-tuning LLMs for CRS are also discussed, highlighting the importance of efficient retrieval methods and the integration of external knowledge for improved system performance.

Contact