|

IBM AI Team Releases Granite 4.0 Nano Series: Compact and Open-Source Small Models Built for AI at the Edge

Small fashions are sometimes blocked by poor instruction tuning, weak software use codecs, and lacking governance. IBM AI crew launched Granite 4.0 Nano, a small mannequin household that targets native and edge inference with enterprise controls and open licensing. The household consists of 8 fashions in two sizes, 350M and about 1B, with each hybrid SSM and transformer variants, every in base and instruct. Granite 4.0 Nano sequence fashions are launched underneath an Apache 2.0 license with native structure assist on widespread runtimes like vLLM, llama.cpp, and MLX

https://huggingface.co/weblog/ibm-granite/granite-4-nano

What is new in Granite 4.0 Nano sequence?

Granite 4.0 Nano consists of 4 mannequin traces and their base counterparts. Granite 4.0 H 1B makes use of a hybrid SSM primarily based structure and is about 1.5B parameters. Granite 4.0 H 350M makes use of the similar hybrid method at 350M. For most runtime portability IBM additionally offers Granite 4.0 1B and Granite 4.0 350M as transformer variations.

Granite launch Sizes in launch Architecture License and governance Key notes
Granite 13B, first watsonx Granite fashions 13B base, 13B instruct, later 13B chat Decoder solely transformer, 8K context IBM enterprise phrases, shopper protections First public Granite fashions for watsonx, curated enterprise knowledge, English focus
Granite Code Models (open) 3B, 8B, 20B, 34B code, base and instruct Decoder solely transformer, 2 stage code coaching on 116 languages Apache 2.0 First totally open Granite line, for code intelligence, paper 2405.04324, out there on HF and GitHub
Granite 3.0 Language Models 2B and 8B, base and instruct Transformer, 128K context for instruct Apache 2.0 Business LLMs for RAG, software use, summarization, shipped on watsonx and HF
Granite 3.1 Language Models (HF) 1B A400M, 3B A800M, 2B, 8B Transformer, 128K context Apache 2.0 Size ladder for enterprise duties, each base and instruct, similar Granite knowledge recipe
Granite 3.2 Language Models (HF) 2B instruct, 8B instruct Transformer, 128K, higher lengthy immediate Apache 2.0 Iterative high quality bump on 3.x, retains enterprise alignment
Granite 3.3 Language Models (HF) 2B base, 2B instruct, 8B base, 8B instruct, all 128K Decoder solely transformer Apache 2.0 Latest 3.x line on HF earlier than 4.0, provides FIM and higher instruction following
Granite 4.0 Language Models 3B micro, 3B H micro, 7B H tiny, 32B H small, plus transformer variants Hybrid Mamba 2 plus transformer for H, pure transformer for compatibility Apache 2.0, ISO 42001, cryptographically signed Start of hybrid era, decrease reminiscence, agent pleasant, similar governance throughout sizes
Granite 4.0 Nano Language Models 1B H, 1B H instruct, 350M H, 350M H instruct, 2B transformer, 2B transformer instruct, 0.4B transformer, 0.4B transformer instruct, complete 8 H fashions are hybrid SSM plus transformer, non H are pure transformer Apache 2.0, ISO 42001, signed, similar 4.0 pipeline Smallest Granite fashions, made for edge, native and browser, run on vLLM, llama.cpp, MLX, watsonx
Table Created by Marktechpost.com

Architecture and coaching

The H variants interleave SSM layers with transformer layers. This hybrid design reduces reminiscence development versus pure consideration, whereas preserving the generality of transformer blocks. The Nano fashions didn’t use a lowered knowledge pipeline. They have been skilled with the similar Granite 4.0 methodology and greater than 15T tokens, then instruction tuned to ship stable software use and instruction following. This carries over strengths from the bigger Granite 4.0 fashions to sub 2B scales.

Benchmarks and aggressive context

IBM compares Granite 4.0 Nano with different underneath 2B fashions, together with Qwen, Gemma, and LiquidAI LFM. Reported aggregates present a major enhance in capabilities throughout normal information, math, code, and security at comparable parameter budgets. On agent duties, the fashions outperform a number of friends on IFEval and on the Berkeley Function Calling Leaderboard v3.

https://huggingface.co/weblog/ibm-granite/granite-4-nano

Key Takeaways

  1. IBM launched 8 Granite 4.0 Nano fashions, 350M and about 1B every, in hybrid SSM and transformer variants, in base and instruct, all underneath Apache 2.0.
  2. The hybrid H fashions, Granite 4.0 H 1B at about 1.5B parameters and Granite 4.0 H 350M at about 350M, reuse the Granite 4.0 coaching recipe on greater than 15T tokens, so functionality is inherited from the bigger household and not a lowered knowledge department.
  3. IBM crew reviews that Granite 4.0 Nano is aggressive with different sub 2B fashions similar to Qwen, Gemma and LiquidAI LFM on normal, math, code and security, and that it outperforms on IFEval and BFCLv3 which matter for software utilizing brokers.
  4. All Granite 4.0 fashions, together with Nano, are cryptographically signed, ISO 42001 licensed and launched for enterprise use, which supplies provenance and governance that typical small group fashions don’t present.
  5. The fashions can be found on Hugging Face and IBM watsonx.ai with runtime assist for vLLM, llama.cpp and MLX, which makes native, edge and browser stage deployments practical for early AI engineers and software program groups.

Editorial Comments

IBM is doing the proper factor right here, it’s taking the similar Granite 4.0 coaching pipeline, the similar 15T token scale, the similar hybrid Mamba 2 plus transformer structure, and pushing it all the way down to 350M and about 1B in order that edge and on system workloads can use the actual governance and provenance story that the bigger Granite fashions have already got. The fashions are Apache 2.0, ISO 42001 aligned, cryptographically signed, and already runnable on vLLM, llama.cpp and MLX. Overall, it is a clear and auditable technique to run small LLMs.


Check out the Model Weights on HF and Technical details. Feel free to take a look at our GitHub Page for Tutorials, Codes and Notebooks. Also, be happy to comply with us on Twitter and don’t neglect to hitch our 100k+ ML SubReddit and Subscribe to our Newsletter. Wait! are you on telegram? now you can join us on telegram as well.

The submit IBM AI Team Releases Granite 4.0 Nano Series: Compact and Open-Source Small Models Built for AI at the Edge appeared first on MarkTechPost.

Similar Posts