Yanqi Yang
I work on LLM alignment and reliability, focusing on reward modeling, tool-use and evaluation.
Research Interests
- Reward modeling. preference learning, reward hacking mitigation and robust reward design under label noise and distribution shift.
- Tool use & agents. Reliable tool-call planning/execution, tool choice & routing.
- Evaluation. Building trustworthy evals for alignment and reliability.