Nemotron-Orchestrator-8B-Qwen3-BF16-NOESIS

BF16 reference checkpoint of nvidia/Nemotron-Orchestrator-8B, losslessly cast from the original FP32 release.

Released as part of the NOESIS Professional Multilingual Dubbing Automation Platform (framework: DHCF-FNO — Deterministic Hybrid Control Framework for Frozen Neural Operators).


⚠️ License notice

This model inherits the NVIDIA Open Model License from the upstream nvidia/Nemotron-Orchestrator-8B. The base model is designated by NVIDIA as "for research and development only".

This BF16 derivative is published as a bandwidth-friendly reference checkpoint for the broader research and development community. Users are responsible for compliance with NVIDIA's license terms — see the LICENSE file in this repository for the full text.


Why this BF16 release exists

The original NVIDIA release ships in FP32 (~32 GB on disk). Most modern inference and quantization tooling (HuggingFace Transformers, vLLM, SGLang, AutoAWQ, AutoGPTQ, llama.cpp BF16 conversion) immediately casts to BF16 on load. Publishing a pre-cast BF16 checkpoint:

  • Halves download bandwidth (16 GB vs 32 GB)
  • Halves disk footprint
  • Skips a slow load-time cast for users
  • Provides a clean BF16 baseline for downstream quantization recipes

The cast is performed via torch.Tensor.to(dtype=torch.bfloat16) with IEEE 754 round-to-nearest-even (PyTorch default). BF16 has the same 8-bit exponent range as FP32 and 7 bits of mantissa, which is lossless for inference-time use of weight tensors.


Model summary

Property Value
Base model nvidia/Nemotron-Orchestrator-8B
Underlying architecture Qwen3-8B (decoder-only transformer, dense, NOT MoE)
Source precision FP32
This release precision BF16
Vocab size 151936
Language English (per base model)
Disk footprint ~16 GB
Inference VRAM ~17 GB BF16 (full-resident on 24 GB+ GPU)

For low-VRAM (6-12 GB) inference, see the AWQ INT4 sibling release: amaimedia/Nemotron-Orchestrator-8B-Qwen3-AWQ-INT4-NOESIS.


How to use

from transformers import AutoModelForCausalLM, AutoTokenizer
import torch

model_id = "amaimedia/Nemotron-Orchestrator-8B-Qwen3-BF16-NOESIS"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
    model_id,
    torch_dtype=torch.bfloat16,
    device_map="auto",
)

prompt = "Plan a multi-step task: find recent AWQ papers, summarize the top three."
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
out = model.generate(**inputs, max_new_tokens=256, do_sample=False)
print(tokenizer.decode(out[0], skip_special_tokens=True))

NOESIS context

This BF16 checkpoint is the source artifact for the AWQ INT4 quantization used as the English orchestration teacher for NOESIS Specialist M9-ORCH-4B during knowledge distillation.

NOESIS is a 9-specialist dubbing automation platform — see the NOESIS collection for the full specialist family.


Acknowledgements & citation

Base model: ToolOrchestra by NVIDIA & University of Hong Kong.

@misc{toolorchestra,
  title  = {ToolOrchestra: Elevating Intelligence via Efficient Model and Tool Orchestration},
  author = {Hongjin Su and Shizhe Diao and Ximing Lu and others},
  year   = {2025},
  eprint = {2511.21689},
  archivePrefix = {arXiv}
}

NOESIS:

@misc{noesis_v14,
  title  = {NOESIS v14.6: DHCF-FNO Multilingual Dubbing Platform},
  author = {Bolotnikov, Ilia},
  year   = {2026},
  publisher = {AMAImedia},
  url    = {https://amaimedia.com}
}
Downloads last month
1,366
Safetensors
Model size
8B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for AMAImedia/Nemotron-Orchestrator-8B-Qwen3-BF16-NOESIS

Finetuned
Qwen/Qwen3-8B
Finetuned
(11)
this model

Paper for AMAImedia/Nemotron-Orchestrator-8B-Qwen3-BF16-NOESIS