0 stringclasses 12 values | 1 float64 0 55.9k |
|---|---|
megatron.core.transformer.attention.forward.qkv | 176.714844 |
megatron.core.transformer.attention.forward.adjust_key_value | 0.100832 |
megatron.core.transformer.attention.forward.rotary_pos_emb | 0.083936 |
megatron.core.transformer.attention.forward.core_attention | 838.296509 |
megatron.core.transformer.attention.forward.linear_proj | 3.306048 |
megatron.core.transformer.transformer_layer._forward_attention.self_attention | 1,020.187622 |
megatron.core.transformer.transformer_layer._forward_attention.self_attn_bda | 234.378403 |
megatron.core.transformer.mlp.forward.linear_fc1 | 0.80816 |
megatron.core.transformer.mlp.forward.activation | 156.001373 |
megatron.core.transformer.mlp.forward.linear_fc2 | 0.775136 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp | 158.650238 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp_bda | 0.37152 |
megatron.core.transformer.attention.forward.qkv | 0.430112 |
megatron.core.transformer.attention.forward.adjust_key_value | 0.07056 |
megatron.core.transformer.attention.forward.rotary_pos_emb | 0.077536 |
megatron.core.transformer.attention.forward.core_attention | 1.235072 |
megatron.core.transformer.attention.forward.linear_proj | 0.332736 |
megatron.core.transformer.transformer_layer._forward_attention.self_attention | 2.566816 |
megatron.core.transformer.transformer_layer._forward_attention.self_attn_bda | 0.30416 |
megatron.core.transformer.mlp.forward.linear_fc1 | 0.35888 |
megatron.core.transformer.mlp.forward.activation | 0.120928 |
megatron.core.transformer.mlp.forward.linear_fc2 | 0.307872 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp | 0.892224 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp_bda | 0.276064 |
megatron.core.transformer.attention.forward.qkv | 171.820038 |
megatron.core.transformer.attention.forward.adjust_key_value | 0.115648 |
megatron.core.transformer.attention.forward.rotary_pos_emb | 0.082944 |
megatron.core.transformer.attention.forward.core_attention | 832.88623 |
megatron.core.transformer.attention.forward.linear_proj | 3.388544 |
megatron.core.transformer.transformer_layer._forward_attention.self_attention | 1,009.967224 |
megatron.core.transformer.transformer_layer._forward_attention.self_attn_bda | 1,004.442322 |
megatron.core.transformer.mlp.forward.linear_fc1 | 3.855808 |
megatron.core.transformer.mlp.forward.activation | 433.085449 |
megatron.core.transformer.mlp.forward.linear_fc2 | 1.371424 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp | 439.813507 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp_bda | 0.310304 |
megatron.core.transformer.attention.forward.qkv | 0.91984 |
megatron.core.transformer.attention.forward.adjust_key_value | 0.071808 |
megatron.core.transformer.attention.forward.rotary_pos_emb | 0.078496 |
megatron.core.transformer.attention.forward.core_attention | 2.30944 |
megatron.core.transformer.attention.forward.linear_proj | 0.181184 |
megatron.core.transformer.transformer_layer._forward_attention.self_attention | 3.83472 |
megatron.core.transformer.transformer_layer._forward_attention.self_attn_bda | 0.065536 |
megatron.core.transformer.mlp.forward.linear_fc1 | 1.01808 |
megatron.core.transformer.mlp.forward.activation | 0.087744 |
megatron.core.transformer.mlp.forward.linear_fc2 | 0.85808 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp | 1.976736 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp_bda | 0.093568 |
megatron.core.transformer.attention.forward.qkv | 173.967911 |
megatron.core.transformer.attention.forward.adjust_key_value | 0.002976 |
megatron.core.transformer.attention.forward.rotary_pos_emb | 0.002976 |
megatron.core.transformer.attention.forward.core_attention | 863.261658 |
megatron.core.transformer.attention.forward.linear_proj | 0.744672 |
megatron.core.transformer.transformer_layer._forward_attention.self_attention | 1,038.817261 |
megatron.core.transformer.transformer_layer._forward_attention.self_attn_bda | 1,058.488403 |
megatron.core.transformer.mlp.forward.linear_fc1 | 5.881504 |
megatron.core.transformer.mlp.forward.activation | 472.481842 |
megatron.core.transformer.mlp.forward.linear_fc2 | 3.470496 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp | 482.59024 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp_bda | 0.23184 |
megatron.core.transformer.attention.forward.qkv | 1.325728 |
megatron.core.transformer.attention.forward.adjust_key_value | 0.002944 |
megatron.core.transformer.attention.forward.rotary_pos_emb | 0.002944 |
megatron.core.transformer.attention.forward.core_attention | 9.87104 |
megatron.core.transformer.attention.forward.linear_proj | 0.698432 |
megatron.core.transformer.transformer_layer._forward_attention.self_attention | 11.919488 |
megatron.core.transformer.transformer_layer._forward_attention.self_attn_bda | 0.231616 |
megatron.core.transformer.mlp.forward.linear_fc1 | 2.960064 |
megatron.core.transformer.mlp.forward.activation | 0.333824 |
megatron.core.transformer.mlp.forward.linear_fc2 | 2.862208 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp | 6.16816 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp_bda | 0.228928 |
megatron.core.transformer.attention.forward.qkv | 171.95401 |
megatron.core.transformer.attention.forward.adjust_key_value | 0.003104 |
megatron.core.transformer.attention.forward.rotary_pos_emb | 0.00304 |
megatron.core.transformer.attention.forward.core_attention | 915.013794 |
megatron.core.transformer.attention.forward.linear_proj | 2.684768 |
megatron.core.transformer.transformer_layer._forward_attention.self_attention | 1,090.517578 |
megatron.core.transformer.transformer_layer._forward_attention.self_attn_bda | 1,033.266602 |
megatron.core.transformer.mlp.forward.linear_fc1 | 14.364896 |
megatron.core.transformer.mlp.forward.activation | 472.968353 |
megatron.core.transformer.mlp.forward.linear_fc2 | 24.896223 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp | 513.041809 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp_bda | 0.896896 |
megatron.core.transformer.attention.forward.qkv | 5.119328 |
megatron.core.transformer.attention.forward.adjust_key_value | 0.00304 |
megatron.core.transformer.attention.forward.rotary_pos_emb | 0.00304 |
megatron.core.transformer.attention.forward.core_attention | 77.115486 |
megatron.core.transformer.attention.forward.linear_proj | 2.659392 |
megatron.core.transformer.transformer_layer._forward_attention.self_attention | 84.919426 |
megatron.core.transformer.transformer_layer._forward_attention.self_attn_bda | 0.902272 |
megatron.core.transformer.mlp.forward.linear_fc1 | 11.422784 |
megatron.core.transformer.mlp.forward.activation | 1.328 |
megatron.core.transformer.mlp.forward.linear_fc2 | 11.5056 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp | 24.268385 |
megatron.core.transformer.transformer_layer._forward_mlp.mlp_bda | 0.902784 |
megatron.core.transformer.attention.forward.qkv | 180.409241 |
megatron.core.transformer.attention.forward.adjust_key_value | 0.111136 |
megatron.core.transformer.attention.forward.rotary_pos_emb | 0.083072 |
megatron.core.transformer.attention.forward.core_attention | 3,018.295898 |
End of preview. Expand
in Data Studio
No dataset card yet
- Downloads last month
- 16