{ "d_model": 1280, "ssm_cfg": { "expand": 1, "d_state": 64 }, "rms_norm_eps": 1e-05, "vocab_size": null, "d_inner": 1280, "d_xb": 1280, "intermediate_size": 5120, "hidden_act": "gelu", "n_layer": 2, "attn_layers": [ 0 ] }