Why Nostr? What is Njump?
2025-04-05 19:59:40

melvincarvalho on Nostr: Maverick and Goose! Llama 4 Maverick: - 17B active parameters, 128 experts, 400B ...

Maverick and Goose!

Llama 4 Maverick:
- 17B active parameters, 128 experts, 400B total.
- 1M token context window.
- Not single-GPU; runs on one H100 DGX host or can be distributed for greater efficiency.
- Outperforms GPT-4o and Gemini 2.0 Flash on coding, reasoning, and multilingual tests at a competitive cost.
- Maintains strong image understanding and grounded reasoning ability.
Author Public Key
npub1melv683fw6n2mvhl5h6dhqd8mqfv3wmxnz4qph83ua4dk4006ezsrt5c24