We've just announced IQ AI.
Jason Wei is an American artificial intelligence researcher known for his contributions to large language models, including the development of chain-of-thought prompting. He is currently a researcher at Meta Superintelligence Labs, having previously worked at OpenAI and Google Brain. [1] [2]
Wei began his career as a research scientist at Google Brain, where his work focused on improving the capabilities of large language models (LLMs). During his time at Google, he was instrumental in research that popularized several key techniques in the field. His work on chain-of-thought prompting demonstrated that LLMs could perform complex reasoning tasks more effectively by generating intermediate steps, mimicking a human-like thought process. He also contributed to the development of instruction tuning, a method for fine-tuning language models to better follow user instructions, and conducted research on the emergent abilities of LLMs, which are unexpected capabilities that appear as models increase in scale. [1] [3]
In February 2023, Wei announced that he had left Google to join the ChatGPT team at OpenAI. [4] At OpenAI, he continued his work on advanced AI models and was a co-creator of the OpenAI o1 model, a series of models designed to spend more time "thinking" before providing an answer to handle more complex problems. He also worked on a project known as "deep research." [1] [2] While at OpenAI, Wei became a strong proponent of reinforcement learning (RL), a training method that uses feedback to refine a model's performance. He described himself as an "RL diehard" and noted how its core concepts influenced his personal philosophy, stating that "beating the teacher requires walking your own path and taking risks and rewards from the environment." [3] [5]
In July 2025, it was reported that Wei, along with his close colleague Hyung Won Chung, was leaving OpenAI to join Meta's newly formed Superintelligence Labs. Wei's expertise in reasoning and reinforcement learning was seen as a key asset for Meta's goals. [1] [2] [6]
Wei has contributed to several influential research areas and models within the field of artificial intelligence. His work has primarily focused on enhancing the reasoning and generalization capabilities of large language models.
While at Google, Wei was a key figure in the research that led to chain-of-thought (CoT) prompting. This technique prompts a language model to break down a multi-step problem into intermediate reasoning steps before giving a final answer. By externalizing its "thought process," the model can arrive at more accurate and reliable solutions for complex tasks in areas like arithmetic, commonsense, and symbolic reasoning. This work, published in a 2022 paper titled "Chain-of-Thought Prompting Elicits Reasoning in Language Models," helped popularize the method and significantly influenced subsequent research in LLM reasoning. [1] [3]
Wei also contributed to research on instruction tuning, particularly with the FLAN (Finetuned Language Net) models at Google. Instruction tuning involves fine-tuning a pre-trained language model on a collection of datasets formatted as instructions. This process improves the model's ability to perform a wide variety of unseen tasks in a zero-shot setting, making it more general-purpose and aligned with user intent. His work is detailed in papers such as "Finetuned Language Models are Zero-Shot Learners" (2021) and "Scaling Instruction-Finetuned Language Models" (2022). Concurrently, he co-authored the paper "Emergent Abilities of Large Language Models" (2022), which characterized how certain capabilities of LLMs appear unpredictably at specific scales of model size and training data, rather than improving smoothly. [1]
At OpenAI, Wei was a co-creator of the o1 model, which was introduced in a preview release in September 2024. The o1 series of models was designed to improve upon the reasoning limitations of previous models by dedicating more computational effort, or "thinking time," before generating a response. This approach allows the model to reason through complex tasks in fields like science, mathematics, and coding with greater accuracy. Wei explained that the model was a significant update to the field of AI, moving beyond simple chain-of-thought prompting to a more sophisticated reasoning process. [7] [2]