Red Teaming Language Model Detectors with Language Models
Z. Shi*, Y. Wang*, F. Yin*, X. Chen, K. Chang, C. Hsieh
TACL, 2023
paper / code
I am a researcher at OpenAI working on posttraining and reinforcement learning. I contributed to GPT-5/GPT-5 thinking, o3/o4-mini, GPT-4o posttraining (e.g., reducing sycophancy), next-gen audio models, improved advanced voice mode, etc.
Previously, I co-founded Converge Lab, an AI startup that aims to bring large language models to the physical world.
I obtained my Ph.D. degree at UCLA CS in 2023, advised by Prof. Cho-Jui Hsieh. Prior to UCLA, I received my B.Eng. in 2019 from the Department of Electronic Engineering, Tsinghua University. I've interned at Google Research and Google DeepMind.
Lion has been successfully deployed in production systems such as Google’s search ads CTR model
Lion has been widely adopted by the community, e.g., MosaicML employed Lion to train their LLMs
Design and source code adapted from Jon Barron’s site