Llama 4 Scout 17B-16E

Meta Llama 4 Scout 17B-16E. 48 decoder layers with interleaved local (RoPE chunked, ×40) and global (NoPE, ×8) attention, MoE FFN (16 experts, top-1). TP=8 on 8×B200 (4×B200×2): 40/8=5 q-heads, 8/8=1 kv-head.

Architecture Overview

Loading visualization...

Architecture Summary

22

Total Modules

10

Blocks

12

Kernels

11

Traced Kernels