Shuo Yang

I am an undergraduate at ACM Honors Class in Shanghai Jiao Tong University. Currently, I am a research assistant at the Sky Computing Lab. I am fortunate to work with Prof. Ion Stoica and become a member of LMSYS. Before that, from 2022 to 2023, I worked with Prof. Tianqi Chen and contributed to TVM.

My research interest is closely related to Large Language Models (LLMs). I am dedicated to making LLMs more user-friendly, efficient, and powerful. Therefore, I am building a low latency inference system for LLM serving, and also exploring directions like LoRA, RAG, and Long Context to extend the capabilities of LLMs.

News

Nov 30, 2023 LLM decontaminator repo receives 250K views, 100 stars and 800 likes!
Nov 14, 2023 Release LLM decontaminator blog!
Aug 1, 2023 OpenAI-compatible API server twitter receives 160K views and 500 likes!
Jun 9, 2023 Release OpenAI-compatible API server blog!

Publications

  1. Rethinking Benchmark and Contamination for Language Models with Rephrased Samples
    Shuo Yang, Wei-Lin Chiang, Lianmin Zheng, and 2 more authors
    2024
  2. S-LoRA: Serving Thousands of Concurrent LoRA Adapters
    Ying Sheng, Shiyi Cao, Dacheng Li, and 9 more authors
    2024