CV

Basics

Name Tianyi Yang
Email m0g1c14n [at] gmail [dot] com
Url https://m0gician.github.io/
Summary I break study everything language model related. Training, inference, optimization, and deployment.

Publications

Projects

  • 2024.08 - Present
    Self-Bootstrap LLM Reasoning Behavior
    Develop a scalable, unsupervised training pipeline for fine-grained LLM reasoning behavior tuning.
    • Developed an unsupervised pipeline using LLM-generated task trees to create self-correcting training data for math and puzzle-solving.
    • Fine-tuned a Qwen2.5 7B base model on the generated dataset to elicit structured, backtracking reasoning.
    • Used RL with custom reward functions for fine-grained control of the model's reasoning, improving efficiency and explainability.
    • Achieved 0.79 on MATH-500, 0.52 on AMC, and 0.20 on AIME24, with generalized reasoning behavior and improved self-correction on downstream tasks.
  • 2025.02 - Present
    Efficient Offline Data Generation Suite for LLMs
    Create a scalable and efficient data generation suite for LLM workflows, now used as the standard for offline data processing at Pinduoduo.
    • Designed a scalable, efficient data generation suite for large-scale offline processing with LLM-powered workflows.
    • Implemented a modular, asynchronous architecture for parallelized data preprocessing, request handling, and storage.
    • Adopted as the standard for offline data generation at Pinduoduo, processing millions of queries weekly across hundreds of millions of products.
  • 2024.02 - Present
    Unsupervised Mass Product Property Extraction
    Developed an unsupervised, config-driven pipeline using LLMs to extract diverse product properties for e-commerce.
    • Developed an unsupervised, config-driven pipeline using LLMs to extract diverse product properties.
    • Used finite-state machines from templates to guide LLM decoding and ensure valid JSON output.
    • Implemented an unsupervised process to clean and filter low-quality or duplicate properties.
    • Built a keyword and feature generation process for billions of products.
    • Designed a holistic evaluation pipeline for property validation, including keyword and semantic checks.
  • 2023.09 - 2024.02
    Red-teaming Benchmark for LLM-Integrated Apps
    Analyzed prompt injection attacks and developed the Raccoon benchmark to evaluate the safety of LLM-integrated applications.
    • Collected 200+ system prompts from applications on OpenAI's GPT Store.
    • Categorized 10+ prompt injection paradigms and constructed compound attacks.
    • Created 20+ templates for in-context defense against prompt injections.
    • Built the Raccoon benchmark to evaluate application safety against complex prompt injection attacks.
  • 2022.02 - 2022.10
    DNN Acceleration via Graph Mutation
    Improved DNN inference efficiency for multi-task systems with a customized graph compiler.
    • Led the design of a parser to convert PyTorch models to an intermediate graph representation.
    • Designed a graph compiler to reuse layers and estimate inference time via sampling.
    • Designed a simulated annealing algorithm to optimize graph merging.
    • Optimized model inference time using TensorRT and TVM.
  • 2021.06 - 2022.01
    Learning Joint Event Relations with Boxes
    Ensure logical consistency across narrative events
    • Designed an XML parser and a data loader for extracting labeled relations and improving the overhead of loading data.
    • Aggregated data with pandas and visualized principal components with existing labels.
    • Designed and implemented experiments that utilizes Longformer to handle larger text input which exceeds RoBERTa's max token size.
  • 2020.09 - 2021.02
    RL Algorithms with High Safety Guarantees
    Developed safe reinforcement learning algorithms with high-confidence policy improvements using the Seldonian Framework.
    • Supervised by Prof. Philip Thomas, I designed and developed machine learning algorithms with high safety constraints using Seldonian Framework.
    • Implemented Importance Sampling estimators for high-confidence policy improvements.
    • Developed a Seldonian Framework library with NumPy and Numba optimizations.
    • Opensourced on GitHub with full documentation.

Work

  • 2024.02 - Present
    NLP Engineer
    Pinduoduo
    Lead LLM-integrated applications for e-commerce scenarios.
    • Lead LLM-integrated applications for e-commerce scenarios.
    • Train and optimize models for latency-critical online tasks.
    • Design and implement LLM development pipelines to optimize model inference performance.
  • 2022.07 - 2023.3
    Software Engineer
    Amazon Robotics
    Develop the next-generation Amazon Grocery Automation System.
    • Designed and implemented AWS cloud infrastructure and algorithms to enable ultra fast 1-2 hour grocery delivery to Amazon customers as part of a new automation pilot program.
    • Led the design of inventory and order language models.
    • Designed an event driven MQTT workflow management system utilizing AWS IoT.
    • Led the test infrastructure design to enable end to end testing of order ingestion to automation fulfillment.
  • 2018.07 - 2018.09
    Research Intern
    Alibaba DAMO Academy
    Research RL-based recommendation systems for Alibaba's online second-hand market Xianyu.
    • Built the homepage Merchant Feed in Xianyu to a classic Contextual Bandit problem.
    • Implemented a recommendation system based on a modified Linear UCB algorithm which utilizes both browsing features and click features.

Education

Awards

Opensource

  • SGLang
    Active contributor: Analyzed DeepSeek V3/R1 inference performance with SGLang optimizations, added support for the Qwen2-57B-A14B model, and improved constrained decoding for Chinese characters.
  • Outlines
    Resolved an exception in FSM construction caused by UTF-8 characters starting with a null byte (\x00).
  • Numba
    Contributed to fixing a faulty UnicodeCharSeq implementation in Numba.

Languages

Chinese Mandarin
Native speaker
English
Fluent
Spanish
Beginner