| --- |
| license: apache-2.0 |
| datasets: |
| - agentica-org/DeepScaleR-Preview-Dataset |
| base_model: |
| - Qwen/Qwen3-4B |
| language: |
| - en |
| pipeline_tag: text-generation |
| library_name: transformers |
| tags: |
| - moe |
| - text-generation-inference |
| - code |
| - deepscale |
| - math |
| --- |
| |
|  |
|
|
| # Segue-Qwen3\_DeepScaleR-Preview |
| |
| > Segue-Qwen3\_DeepScaleR-Preview is an experimental fine-tuned variant of the Qwen3-4B model architecture. It is trained on the DeepScaleR-Preview dataset—comprising high-quality mathematical reasoning problems—to achieve exceptional performance in symbolic, mathematical, and logical tasks with lightweight computational requirements. |
|
|
| ## Key Features |
|
|
| 1. Precision Reasoning with DeepScaleR-Preview Dataset |
| Fine-tuned on approximately 40,000 curated math problem-answer pairs sourced from: |
|
|
| * AIME (1984–2023) |
| * AMC (pre-2023) |
| * Omni-MATH |
| This enables superior symbolic manipulation and step-by-step logical deduction. |
| |
| 2. Lightweight Code Understanding |
| Capable of interpreting and generating correct code in Python, C++, and other logic-intensive languages with an emphasis on problem-solving and structured thought. |
|
|
| 3. Structured Output Formatting |
| Outputs are designed to be well-formatted in Markdown, JSON, LaTeX, or tables—ideal for technical documentation, math notebooks, and data workflows. |
|
|
| 4. Instruction-Following Accuracy |
| Strong multi-step instruction adherence, particularly for STEM domains. Ensures continuity, factual correctness, and process transparency in reasoning chains. |
|
|
| 5. Multilingual Capabilities |
| Supports over 20 languages for mathematical and logical reasoning, technical instruction translation, and cross-lingual academic support. |
|
|
| 6. Efficient 4B Architecture |
| Built on the Qwen3-4B base model to balance performance and scalability. Runs efficiently on mid-range GPUs while delivering high-accuracy inference. |
|
|
| ## Quickstart with Transformers |
|
|
| ```python |
| from transformers import AutoModelForCausalLM, AutoTokenizer |
| |
| model_name = "prithivMLmods/Segue-Qwen3_DeepScaleR-Preview" |
| |
| model = AutoModelForCausalLM.from_pretrained( |
| model_name, |
| torch_dtype="auto", |
| device_map="auto" |
| ) |
| tokenizer = AutoTokenizer.from_pretrained(model_name) |
| |
| prompt = "Solve for x: 5(x - 2) = 3x + 4, showing all steps clearly." |
| |
| messages = [ |
| {"role": "system", "content": "You are a precise mathematical assistant trained on DeepScaleR-Preview dataset."}, |
| {"role": "user", "content": prompt} |
| ] |
| |
| text = tokenizer.apply_chat_template( |
| messages, |
| tokenize=False, |
| add_generation_prompt=True |
| ) |
| |
| model_inputs = tokenizer([text], return_tensors="pt").to(model.device) |
| |
| generated_ids = model.generate( |
| **model_inputs, |
| max_new_tokens=512 |
| ) |
| generated_ids = [ |
| output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids) |
| ] |
| |
| response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0] |
| print(response) |
| ``` |
|
|
| ## Intended Use |
|
|
| * Step-by-step mathematical problem solving |
| * Symbolic computation and logic derivation |
| * Code generation and correction in technical environments |
| * Automated LaTeX/Markdown/JSON generation for education and documentation |
| * Academic tutoring and educational assistants |
| * Multilingual reasoning and translation of structured content |
|
|
| ## Limitations |
|
|
| * Less suitable for open-domain conversation or creative writing |
| * Smaller context window compared to large-scale LLMs |
| * May be sensitive to token formatting in edge-case symbolic prompts |
| * Could underperform on intentionally adversarial logic inputs |
|
|
| ## References |
|
|
| 1. Qwen2.5 Technical Report – [https://arxiv.org/pdf/2412.15115](https://arxiv.org/pdf/2412.15115) |
| 2. YaRN: Context Window Extension for LLMs – [https://arxiv.org/pdf/2309.00071](https://arxiv.org/pdf/2309.00071) |