4. Model-by-model analysis
Each section includes: (1) the axis scores used in the composite, (2) a plain-language rationale,
and (3) direct links to primary artifacts (model cards, licenses, blogs, technical reports, and deployment recipes) embedded inline.
#1 Trinity Mini (Arcee Trinity family) — Composite 93.0
Open-weight MoE designed for local/VPC deployment with unusually strong base-checkpoint availability and documentation.
- Local-first reality: Trinity Mini is a 26B MoE with ~3B active; a GGUF distribution exists for practical local inference.
- Base access: explicit base checkpoint (Trinity‑Mini‑Base) in addition to the instruction checkpoint (Trinity Mini).
- Technical transparency: family-level technical report with architecture and training details (Arcee Trinity Large Technical Report).
- License clarity: Apache‑2.0 license stated on model card artifacts (e.g., Base, GGUF).
#2 Ministral 3B — Composite 85.5
Edge-first dense model with vision and large context in the Ministral 3 line; strong “everywhere deployment” profile.
#3 Jamba2 3B — Composite 84.75
On-device, long-context hybrid architecture tuned for reliable instruction following with Apache‑2.0 open weights.
- Release details: AI21 introduces the Jamba2 open-source family (Apache‑2.0; 256K context) (AI21 release blog).
- Concrete anchor: Jamba2 3B model card emphasizes on-device deployment and reliability.
- Docs: AI21 documentation covers Jamba2 Mini and Jamba2 3B roles and positioning (AI21 Jamba docs).
#4 Ministral 8B — Composite 84.4
Balanced 8B dense model: strong capability with local viability; anchored to the official Instruct checkpoint.
#5 Ministral 14B — Composite 82.7
High-end dense member of Ministral 3: better reasoning and breadth at a higher local cost.
#6 Jamba2 Mini — Composite 81.45
MoE member of the Jamba2 family; higher capability than 3B at the cost of heavier deployment requirements.
- Anchor checkpoint: AI21‑Jamba2‑Mini (Apache‑2.0).
- Release facts: AI21 lists family fast facts (Apache‑2.0, model sizes, 256K context) (AI21 blog).
- Docs: model positioning and benefits in AI21 docs (Jamba docs).
#7 Phi‑4 — Composite 80.5
MIT-licensed small model with strong reasoning for size; high local viability; more corporate safety post-training.
#8 Mistral Large 3 — Composite 76.75
Frontier-scale open MoE (41B active / 675B total) under Apache‑2.0; exceptional capability with datacenter-level hardware needs.
#9 IBM Granite 4.0 Micro — Composite 75.55
Apache‑2.0 open-weight enterprise-oriented family with strong docs and edge-friendly variants.
- Anchor checkpoint: granite‑4.0‑micro (Apache‑2.0).
- Docs: Granite 4.0 documentation highlights edge workflows and Apache‑2.0 (IBM Granite docs).
- Repo: Granite 4.0 language models repository (Apache‑2.0) (GitHub repo).
- Base variant: micro base card includes training strategy description (micro‑base).
#10 Snowflake Arctic — Composite 71.4
Open Apache‑2.0 enterprise MoE hybrid with strong documentation and a clear focus on data/SQL workflows.
#11 Step‑3.5‑Flash — Composite 71.0
Open Apache‑2.0 MoE (~196B total / ~11B active) built for agentic workloads; high capability, discounted on sovereignty.
- Model card: Step‑3.5‑Flash (Apache‑2.0; BF16 weights; ~199B params).
- Code repo: architecture overview and “11B active” framing (GitHub repo).
- Technical paper: arXiv 2602.10604 describing the RL framework and eval results.
- Base checkpoint access: mid-train base is published (Base‑Midtrain).
#12 DeepSeek‑R1 — Composite 70.55
MIT-licensed open reasoning model series; strong capability and modifiability, discounted for stack-entanglement risk.
- Model card: DeepSeek‑R1 explicitly states MIT licensing and derivative rights.
- License file: MIT license text in-repo (LICENSE).
- Official repo: DeepSeek’s GitHub release notes include distillation/derivative permissions (GitHub repo).
- Policy pressure example: Reuters reporting on censorship-focused derivatives illustrates geopolitical alignment pressure (Reuters: R1‑Safe).
#13 Qwen 3.5 SLM family — Composite 68.3
Small open weights with strong performance and extensive tooling; discounted on sovereignty due to deep platform entanglement.
- Official blog: Qwen3.5 “Towards Native Multimodal Agents” (Qwen blog).
- License statement: QwenLM states all open-weight Qwen3.5 models are Apache‑2.0 (GitHub: Qwen3.5).
- Anchor small checkpoints: 0.8B, 2B, 4B, 9B.
- Local packaging: Ollama model page indicates easy local pulls for small variants (Ollama: qwen3.5:2b).
#14 GLM‑5 — Composite 65.5
Frontier-scale open-weight MoE (up to 744B total / 40B active) with MIT licensing; datacenter hardware profile.
- Model card: zai‑org/GLM‑5.
- Official blog: GLM‑5 release notes state weights are released under MIT (z.ai blog).
- Repo: GLM‑5 GitHub repository with architecture framing (GitHub: GLM‑5).
#15 DeepSeek‑V3.2 — Composite 65.25
MIT-licensed reasoning-first successor to V3; frontier-level capability with datacenter deployment characteristics.
#16 DeepSeek‑V3.2‑Exp — Composite 65.0
Experimental sparse-attention version; strong capability, similar sovereignty profile to V3.2 with long-context efficiency focus.
#17 Seed‑OSS‑36B — Composite 64.5
Apache‑2.0 open dense 36B with long context and strong reasoning; penalized for heavier hardware and strong platform entanglement.
End of model cards. The page deliberately embeds primary links within each model section to avoid a “links-only appendix”.