Xiaoquan KongAI Engineer & Researcher I build AI systems that work at scale—from research prototypes to production platforms serving millions. My work focuses on agentic AI, retrieval-augmented generation, and reinforcement learning. At Baidu and Alibaba, I built conversational AI systems deployed across millions of vehicles serving large-scale production workloads. I hold a MEng from Duke University and am a Google Developer Expert in Machine Learning. Email: u1mail2me@gmail.com / GitHub / LinkedIn / Google Scholar |
|
Reinforcement Learning Curriculum Development [course site] [course]
Duke University
Designed curriculum for graduate RL course (AIPI 590) covering tabular methods to modern RLHF. Built complete alignment pipeline from scratch—reward modeling, PPO training, safety evaluation—demonstrating significant improvements in model safety. Students implement core algorithms from first principles, learning to align language models with human preferences.
Educational AI Development
Duke CREATE (Center for Research & Engineering of AI Technology in Education)
Developing QUBIT, an AI programming assistant that teaches through explanation rather than direct code generation. Implementing scaffolded prompting and constraint-based hints to guide students through problem-solving while maintaining engagement.
A minimal implementation (~300 lines) exposing the complete control flow of agentic systems. By avoiding framework abstractions, it reveals how tool-calling and multi-agent coordination actually work—enabling developers to understand and modify agent behaviors at the implementation level. Supports major LLMs via LiteLLM.
PyTorch-compatible ML framework built from scratch—autodiff engine, neural networks, optimizers. Designed for understanding deep learning fundamentals with drop-in replacement capability.
Enterprise-grade RAG processing text, images, videos, and audio with evidence-based answer generation. Built with Vertex AI, Pinecone, and Gemini-1.5-Pro for auditable decision-making.
Suite of Chinese NLP tools including spaCy models, character decomposition library, and educational tokenizer. Widely adopted in industry and academia with strong community engagement.
550+ citations, h-index: 7 • [Google Scholar]
* denotes equal contribution
Duke University