Picture for Aohan Zeng

Aohan Zeng

Data-Efficient RLVR via Off-Policy Influence Guidance

Add code
Oct 30, 2025
Viaarxiv icon

GLM-4.5: Agentic, Reasoning, and Coding (ARC) Foundation Models

Add code
Aug 08, 2025
Viaarxiv icon

GLM-4.1V-Thinking: Towards Versatile Multimodal Reasoning with Scalable Reinforcement Learning

Add code
Jul 02, 2025
Figure 1 for GLM-4.1V-Thinking: Towards Versatile Multimodal Reasoning with Scalable Reinforcement Learning
Figure 2 for GLM-4.1V-Thinking: Towards Versatile Multimodal Reasoning with Scalable Reinforcement Learning
Figure 3 for GLM-4.1V-Thinking: Towards Versatile Multimodal Reasoning with Scalable Reinforcement Learning
Figure 4 for GLM-4.1V-Thinking: Towards Versatile Multimodal Reasoning with Scalable Reinforcement Learning
Viaarxiv icon

Does RLHF Scale? Exploring the Impacts From Data, Model, and Method

Add code
Dec 08, 2024
Figure 1 for Does RLHF Scale? Exploring the Impacts From Data, Model, and Method
Figure 2 for Does RLHF Scale? Exploring the Impacts From Data, Model, and Method
Figure 3 for Does RLHF Scale? Exploring the Impacts From Data, Model, and Method
Figure 4 for Does RLHF Scale? Exploring the Impacts From Data, Model, and Method
Viaarxiv icon

GLM-4-Voice: Towards Intelligent and Human-Like End-to-End Spoken Chatbot

Add code
Dec 03, 2024
Viaarxiv icon

Scaling Speech-Text Pre-training with Synthetic Interleaved Data

Add code
Nov 26, 2024
Figure 1 for Scaling Speech-Text Pre-training with Synthetic Interleaved Data
Figure 2 for Scaling Speech-Text Pre-training with Synthetic Interleaved Data
Figure 3 for Scaling Speech-Text Pre-training with Synthetic Interleaved Data
Figure 4 for Scaling Speech-Text Pre-training with Synthetic Interleaved Data
Viaarxiv icon

VisualAgentBench: Towards Large Multimodal Models as Visual Foundation Agents

Add code
Aug 12, 2024
Figure 1 for VisualAgentBench: Towards Large Multimodal Models as Visual Foundation Agents
Figure 2 for VisualAgentBench: Towards Large Multimodal Models as Visual Foundation Agents
Figure 3 for VisualAgentBench: Towards Large Multimodal Models as Visual Foundation Agents
Figure 4 for VisualAgentBench: Towards Large Multimodal Models as Visual Foundation Agents
Viaarxiv icon

ChatGLM: A Family of Large Language Models from GLM-130B to GLM-4 All Tools

Add code
Jun 18, 2024
Figure 1 for ChatGLM: A Family of Large Language Models from GLM-130B to GLM-4 All Tools
Figure 2 for ChatGLM: A Family of Large Language Models from GLM-130B to GLM-4 All Tools
Figure 3 for ChatGLM: A Family of Large Language Models from GLM-130B to GLM-4 All Tools
Figure 4 for ChatGLM: A Family of Large Language Models from GLM-130B to GLM-4 All Tools
Viaarxiv icon

ChatGLM-Math: Improving Math Problem-Solving in Large Language Models with a Self-Critique Pipeline

Add code
Apr 03, 2024
Figure 1 for ChatGLM-Math: Improving Math Problem-Solving in Large Language Models with a Self-Critique Pipeline
Figure 2 for ChatGLM-Math: Improving Math Problem-Solving in Large Language Models with a Self-Critique Pipeline
Figure 3 for ChatGLM-Math: Improving Math Problem-Solving in Large Language Models with a Self-Critique Pipeline
Figure 4 for ChatGLM-Math: Improving Math Problem-Solving in Large Language Models with a Self-Critique Pipeline
Viaarxiv icon

ChatGLM-RLHF: Practices of Aligning Large Language Models with Human Feedback

Add code
Apr 03, 2024
Figure 1 for ChatGLM-RLHF: Practices of Aligning Large Language Models with Human Feedback
Figure 2 for ChatGLM-RLHF: Practices of Aligning Large Language Models with Human Feedback
Figure 3 for ChatGLM-RLHF: Practices of Aligning Large Language Models with Human Feedback
Figure 4 for ChatGLM-RLHF: Practices of Aligning Large Language Models with Human Feedback
Viaarxiv icon