About 34,400 results
Open links in new tab
  1. [New Model]: Qwen2.5-Math-PRM-7B, Qwen2.5-Math-PRM-72B, and …

    Jan 17, 2025 · [New Model]: Qwen2.5-Math-PRM-7B, Qwen2.5-Math-PRM-72B, and Qwen2.5-Math-7B-PRM800K #12140 Closed Lolo1222 opened on Jan 16

  2. [RFC]: Let every model be a reward model/embedding model for PRMs

    Oct 12, 2024 · To support this kind of PRMs, we can simply add an default pooler method to all models (because we can see that math-shepherd-mistral-7b-prm is just a LlamaForCausalLM) and allow …

  3. [Installation]: with latest vllm source code installation done, but ...

    Jul 31, 2025 · the error log is here. it said flashinfer failed to find C++ math.h, but I am pretty sure the file is there by ls

  4. Comparing d9c8f8c...2fc05d5 · vllm-project/vllm · GitHub

    Nov 25, 2025 · A high-throughput and memory-efficient inference and serving engine for LLMs - Comparing d9c8f8c...2fc05d5 · vllm-project/vllm

  5. [New Model]: Qwen2.5-Math-PRM-7B, Qwen2.5-Math-PRM-72B …

    Jan 19, 2025 · Never mind, num_labels is defined correctly after loading AutoConfig. Actually num_labels is set to 2 for both Qwen/Qwen2.5-Math-RM-72B and Qwen2.5-Math-PRM-7B, so we …

  6. [Usage] Qwen3 Usage Guide · Issue #17327 · vllm-project/vllm - GitHub

    Apr 28, 2025 · vLLM v0.8.4 and higher natively supports all Qwen3 and Qwen3MoE models. Example command: vllm serve Qwen/... --enable-reasoning --reasoning-parser deepseek_r1 All models …

  7. vllm/vllm/model_executor/models/minicpmv.py at main - GitHub

    import math from collections import defaultdict from collections.abc import Callable, Iterable, Mapping, Sequence from functools import partial

  8. [Bug]: Qwen2.5-Math-7B-Instruct vllm output garbled code, but the ...

    Oct 9, 2024 · Model Input Dumps No response 🐛 Describe the bug Hi! I'm now using Qwen2.5-Math-7B-Instruct to solve problems in the MATH dataset. And I found that the vLLM engine sometimes has …

  9. [Bug]: The output of MathResponse is empty when running THUDM

    Apr 22, 2025 · class MathResponse (BaseModel): steps: list [Step] final_answer: str client = OpenAI (base_url=" [http://localhost:8000/v1",api_key="-] (http://localhost:8000/v1%22,api_key=%22-)") …

  10. Comparing 4b52e64...59f3541 · vllm-project/vllm · GitHub

    A high-throughput and memory-efficient inference and serving engine for LLMs - Comparing 4b52e64...59f3541 · vllm-project/vllm