
[New Model]: Qwen2.5-Math-PRM-7B, Qwen2.5-Math-PRM-72B, …
Jan 17, 2025 · [New Model]: Qwen2.5-Math-PRM-7B, Qwen2.5-Math-PRM-72B, and Qwen2.5-Math-7B-PRM800K #12140 Closed Lolo1222 opened on Jan 16
[RFC]: Let every model be a reward model/embedding model for …
Oct 12, 2024 · To support this kind of PRMs, we can simply add an default pooler method to all models (because we can see that math-shepherd-mistral-7b-prm is just a LlamaForCausalLM) …
[Installation]: with latest vllm source code installation done, but ...
Jul 31, 2025 · the error log is here. it said flashinfer failed to find C++ math.h, but I am pretty sure the file is there by ls
Comparing d9c8f8c...2fc05d5 · vllm-project/vllm · GitHub
Nov 25, 2025 · A high-throughput and memory-efficient inference and serving engine for LLMs - Comparing d9c8f8c...2fc05d5 · vllm-project/vllm
[New Model]: Qwen2.5-Math-PRM-7B, Qwen2.5-Math-PRM-72B …
Jan 19, 2025 · Never mind, num_labels is defined correctly after loading AutoConfig. Actually num_labels is set to 2 for both Qwen/Qwen2.5-Math-RM-72B and Qwen2.5-Math-PRM-7B, so …
[Usage] Qwen3 Usage Guide · Issue #17327 · vllm-project/vllm
Apr 28, 2025 · vLLM v0.8.4 and higher natively supports all Qwen3 and Qwen3MoE models. Example command: vllm serve Qwen/... --enable-reasoning --reasoning-parser deepseek_r1 …
vllm/vllm/model_executor/models/minicpmv.py at main - GitHub
import math from collections import defaultdict from collections.abc import Callable, Iterable, Mapping, Sequence from functools import partial
[Bug]: Qwen2.5-Math-7B-Instruct vllm output garbled code, but …
Oct 9, 2024 · Model Input Dumps No response 🐛 Describe the bug Hi! I'm now using Qwen2.5-Math-7B-Instruct to solve problems in the MATH dataset. And I found that the vLLM engine …
[Bug]: The output of MathResponse is empty when running …
Apr 22, 2025 · class MathResponse (BaseModel): steps: list [Step] final_answer: str client = OpenAI (base_url=" [http://localhost:8000/v1",api_key="-] …
Comparing 4b52e64...59f3541 · vllm-project/vllm · GitHub
A high-throughput and memory-efficient inference and serving engine for LLMs - Comparing 4b52e64...59f3541 · vllm-project/vllm