Siyuan (Wilson) Sun孙思源

Undergraduate Student @ University of Arizona 人工智能/计算机科学本科生 @ 亚利桑那大学

About Me关于我

I am an undergraduate student at University of Arizona, majoring in AI/CS. Currently, I am conducting Natural Language Processing (NLP) research at CLULab and advised by Prof. Mihai Surdeanu. 我是就读于亚利桑那大学的本科三年级学生,主修人工智能/计算机科学。目前,我主要在CLULab进行自然语言处理(NLP)研究,有幸接受Mihai Surdeanu教授的指导。

My general research interests lie at the intersection of Large Language Models (LLMs) and Information Retrieval (IR), with a specific focus on Memory Mechanisms, Model Alignment, and the application of Reinforcement Learning in NLP. 我的研究兴趣集中在大语言模型 (LLMs)信息检索 (IR) 的交叉领域,重点关注记忆机制模型对齐以及强化学习理论在NLP中的应用。

I value research that is both intellectually profound and practically impactful. I specialize in leveraging rigorous, large-scale experimentation to navigate and refine my research directions. 我对学术研究有深入的思考,同时很注重研究成果的实际应用价值。我擅长运用严谨的大规模实验来探索并优化我的研究方向。

I am actively seeking Internship Opportunities in both academia and industry. I am also preparing for Fall 2027 Ph.D./Master applications in NLP and Machine Learning. If you have an open position that aligns with my profile, I would be delighted to connect. 🤝 我正积极寻求学术界及工业界的实习机会。同时,我也在寻找2027年秋季入学,NLP与机器学习领域的博士或硕士项目机会。如果您有与我的背景相契合的职位空缺,我将非常乐意与您取得联系。🤝

Education教育经历

B.S. in Artificial Intelligence 人工智能 理学学士

Aug 2025 - May 2027 (Expected) 2025年8月 - 2027年5月 (预期)
University of Arizona 亚利桑那大学
Tucson, AZ 美国 亚利桑那州
  • GPA: 4.0/4.0 GPA: 4.0/4.0
  • Relevant Coursework: Deep Learning for NLP, Neural Networks, Reinforcement Learning, Principles of ML, Text Retrieval and Web Search, etc. 相关课程:自然语言处理深度学习、神经网络、强化学习、机器学习原理、文本检索与网络搜索等。

B.S. in Artificial Intelligence 人工智能 理学学士

Sep 2023 - July 2025 2023年9月 - 2025年7月
East China University of Science and Technology 华东理工大学
Shanghai, China 中国 上海
  • Transferred to University of Arizona 转学至亚利桑那大学

Experience专业经历

Undergraduate Research Assistant 本科生研究助理

Sep 2025 - Present 2025年9月 - 至今
CLULab | Advised by Prof. Mihai Surdeanu CLULab | 导师:Mihai Surdeanu 教授
Tucson, AZ 美国 亚利桑那州
  • Generalization-aware optimization of Transformer-based Language Models Transformer架构的语言模型可泛化调优
  • Architectural Innovations in Neural Retrieval Systems 以神经网络为基础的信息检索架构创新

Technical Skills & Research Competencies 技术技能与科研能力

Deep Learning & NLP Research 深度学习与自然语言处理科研

Model Alignment: SFT, RLHF (PPO/DPO), Reward-Guided Alignment, EM-based optimization for generative models.
Transformer-based Language Modeling: LLMs (Qwen, Llama, RoBERTa, GPT), Bi-Encoder, Cross-Encoder, PEFT (LoRA/QLoRA).
Neural IR: Dense Retrieval, Query Expansion, Reranking, End-to-end Joint Optimization of Retrieval Pipelines.
模型对齐: 有监督微调 (SFT)、强化学习对齐 (PPO/DPO)、奖励引导对齐、基于 EM 算法的生成模型优化。
语言建模: 大语言模型 (Qwen, RoBERTa, GPT)、参数高效微调 (PEFT/LoRA/QLoRA)、Transformer 架构调优。
神经信息检索: 稠密检索、查询扩展 (QE)、交叉编码重排序、检索流水线的端到端联合优化。

ML Engineering & Infrastructure 机器学习工程与基础设施

Frameworks: PyTorch, Hugging Face (Transformers, Accelerate, PEFT, TRL), Scikit-learn, LangChain.
Vector Databases & IR: FAISS (HNSW/IVF), Milvus, Elasticsearch/Lucene (BM25), Open-source Foundations (BGE, Jina).
Optimization: Multi-GPU Parallel Training, Mixed Precision Training (FP16/BF16), CUDA-level memory management (OOM handling).
框架: PyTorch, Hugging Face (Transformers, Accelerate, PEFT, TRL), Scikit-learn, LangChain。
向量数据库与检索: FAISS (HNSW/IVF), Milvus, Elasticsearch/Lucene (BM25), 开源检索基础模型 (BGE, Jina)。
性能优化: 多显卡并行训练、混合精度训练 (FP16/BF16)、CUDA 显存管理与大规模实验 OOM 调优。

Programming & System Arch 编程语言与系统架构

Python, Java, C/C++, Shell, Linux (remote development), Git, Gymnasium (Reinforcement Learning environments). Python, Java, C/C++, Shell 脚本, Linux 环境远程开发, Git, Gymnasium (强化学习环境构建)。