HyperAI

vLLM+Open WebUI Deploys AM-Thinking-v1 Dense Language Model

1. Tutorial Introduction

AM-Thinking-v1 is a 32 B dense language model developed by am-team on May 10, 2025, focusing on enhancing reasoning capabilities. The model shows strong performance in reasoning benchmarks, comparable to large MoE models such as DeepSeek-R1, Qwen3-235B-A22B, Seed1.5-Thinking, and larger dense models such as Nemotron-Ultra-253B-v1. The related paper results are "AM-Thinking-v1: Advancing the Frontier of Reasoning at 32B Scale".

This tutorial uses the resources for the dual-SIM A6000.

2. Project Examples

3. Operation steps

1. After starting the container, click the API address to enter the Web interface

If "Model" is not displayed, it means the model is being initialized. Since the model is large, please wait about 1-2 minutes and refresh the page.

2. After entering the webpage, you can start a conversation with the model

How to use

4. Discussion

🖌️ If you see a high-quality project, please leave a message in the background to recommend it! In addition, we have also established a tutorial exchange group. Welcome friends to scan the QR code and remark [SD Tutorial] to join the group to discuss various technical issues and share application effects↓

Citation Information

The citation information for this project is as follows:

@misc{ji2025amthinkingv1advancingfrontierreasoning,
      title={AM-Thinking-v1: Advancing the Frontier of Reasoning at 32B Scale}, 
      author={Yunjie Ji and Xiaoyu Tian and Sitong Zhao and Haotian Wang and Shuaiting Chen and Yiping Peng and Han Zhao and Xiangang Li},
      year={2025},
      eprint={2505.08311},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2505.08311}, 
}