Support our open-source dataset and model releases!

image/jpeg

Shining Valiant 3: Qwen3-1.7B, Qwen3-4B, Qwen3-8B, Ministral-3-14B-Reasoning-2512, gpt-oss-20b

Shining Valiant 3 is a science, AI design, and general reasoning specialist built on Ministral 3.

Prompting Guide

Shining Valiant 3 uses the Ministral-3-14B-Reasoning-2512 prompt format.

Example inference script to get started:

import torch
from transformers import Mistral3ForConditionalGeneration, MistralCommonBackend

model_id = "ValiantLabs/Ministral-3-14B-Reasoning-2512-ShiningValiant3"

tokenizer = MistralCommonBackend.from_pretrained(model_id)

model = Mistral3ForConditionalGeneration.from_pretrained(
    model_id, torch_dtype=torch.bfloat16, device_map="auto"
)

user_prompt = "Propose a novel cognitive architecture where the primary memory component is a Graph Neural Network (GNN). How would this GNN represent working, declarative, and procedural memory? How would the \"cognitive cycle\" be implemented as operations on this graph?"

system_prompt = (
    "# HOW YOU SHOULD THINK AND ANSWER\n\n"
    "First draft your thinking process (inner monologue) until you arrive at a response. "
    "Format your response using Markdown, and use LaTeX for any mathematical equations. "
    "Write both your thoughts and the response in the same language as the input.\n\n"
    "Your thinking process must follow the template below:"
    "[THINK]Your thoughts or/and draft, like working through an exercise on scratch paper. "
    "Be as casual and as long as you want until you are confident to generate the response to the user.[/THINK]"
    "Here, provide a self-contained response."
)

messages = [
    {
        "role": "system",
        "content": system_prompt
    },
    {
        "role": "user",
        "content": [
            {
                "type": "text",
                "text": user_prompt,
            },
        ],
    },
]

tokenized = tokenizer.apply_chat_template(messages, return_tensors="pt", return_dict=True)
tokenized = {k: v.to("cuda") for k, v in tokenized.items() if hasattr(v, "to")}

output = model.generate(
    **tokenized,
    max_new_tokens=20000,
)[0]

decoded_output = tokenizer.decode(output[len(tokenized["input_ids"][0]):])
print(decoded_output)

image/jpeg

Shining Valiant 3 is created by Valiant Labs.

Check out our HuggingFace page to see all of our models!

We care about open source. For everyone to use.

Downloads last month
23
Safetensors
Model size
14B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for ValiantLabs/Ministral-3-14B-Reasoning-2512-ShiningValiant3

Finetuned
(7)
this model
Merges
1 model
Quantizations
2 models

Datasets used to train ValiantLabs/Ministral-3-14B-Reasoning-2512-ShiningValiant3

Collection including ValiantLabs/Ministral-3-14B-Reasoning-2512-ShiningValiant3