Skip to content

Commit 2955061

Browse files
localai-botmudler
andauthored
chore(model gallery): 🤖 add 1 new models via gallery agent (#6910)
chore(model gallery): 🤖 add new models via gallery agent Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com> Co-authored-by: mudler <[email protected]>
1 parent 84644ab commit 2955061

File tree

1 file changed

+37
-0
lines changed

1 file changed

+37
-0
lines changed

gallery/index.yaml

Lines changed: 37 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -23010,3 +23010,40 @@
2301023010
- filename: Qwen3-VLTO-32B-Instruct.i1-Q4_K_S.gguf
2301123011
sha256: 789d55249614cd1acee1a23278133cd56ca898472259fa2261f77d65ed7f8367
2301223012
uri: huggingface://mradermacher/Qwen3-VLTO-32B-Instruct-i1-GGUF/Qwen3-VLTO-32B-Instruct.i1-Q4_K_S.gguf
23013+
- !!merge <<: *qwen3
23014+
name: "qwen3-vlto-32b-thinking"
23015+
urls:
23016+
- https://huggingface.co/mradermacher/Qwen3-VLTO-32B-Thinking-GGUF
23017+
description: |
23018+
**Model Name:** Qwen3-VLTO-32B-Thinking
23019+
**Model Type:** Large Language Model (Text-Only)
23020+
**Base Model:** Qwen/Qwen3-VL-32B-Thinking (vanilla Qwen3-VL-32B with vision components removed)
23021+
**Architecture:** Transformer-based, 32-billion parameter model optimized for reasoning and complex text generation.
23022+
23023+
### Description:
23024+
Qwen3-VLTO-32B-Thinking is a pure text-only variant of the Qwen3-VL-32B-Thinking model, stripped of its vision capabilities while preserving the full reasoning and language understanding power. It is derived by transferring the weights from the vision-language model into a text-only transformer architecture, maintaining the same high-quality behavior for tasks such as logical reasoning, code generation, and dialogue.
23025+
23026+
This model is ideal for applications requiring deep linguistic reasoning and long-context understanding without image input. It supports advanced multimodal reasoning capabilities *in text form*—perfect for research, chatbots, and content generation.
23027+
23028+
### Key Features:
23029+
- ✅ 32B parameters, high reasoning capability
23030+
- ✅ No vision components — fully text-only
23031+
- ✅ Trained for complex thinking and step-by-step reasoning
23032+
- ✅ Compatible with Hugging Face Transformers and GGUF inference tools
23033+
- ✅ Available in multiple quantization levels (Q2_K to Q8_0) for efficient deployment
23034+
23035+
### Use Case:
23036+
Ideal for advanced text generation, logical inference, coding, and conversational AI where vision is not needed.
23037+
23038+
> 🔗 **Base Model**: [Qwen/Qwen3-VL-32B-Thinking](https://huggingface.co/Qwen/Qwen3-VL-32B-Thinking)
23039+
> 📦 **Quantized Versions**: Available via [mradermacher/Qwen3-VLTO-32B-Thinking-GGUF](https://huggingface.co/mradermacher/Qwen3-VLTO-32B-Thinking-GGUF)
23040+
23041+
---
23042+
*Note: The original model was created by Alibaba’s Qwen team. This variant was adapted by qingy2024 and quantized by mradermacher.*
23043+
overrides:
23044+
parameters:
23045+
model: Qwen3-VLTO-32B-Thinking.Q4_K_M.gguf
23046+
files:
23047+
- filename: Qwen3-VLTO-32B-Thinking.Q4_K_M.gguf
23048+
sha256: d88b75df7c40455dfa21ded23c8b25463a8d58418bb6296304052b7e70e96954
23049+
uri: huggingface://mradermacher/Qwen3-VLTO-32B-Thinking-GGUF/Qwen3-VLTO-32B-Thinking.Q4_K_M.gguf

0 commit comments

Comments
 (0)