Skip to content

Commit e9d0051

Browse files
authored
add llama3.1-8b config (#126)
* add llama3.1-8b config * nit
1 parent 00d5f39 commit e9d0051

File tree

1 file changed

+27
-0
lines changed

1 file changed

+27
-0
lines changed
Lines changed: 27 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,27 @@
1+
defaults:
2+
- _self_ # refers to this config file
3+
- scaling: llama-fsdp # refers to scaling/llama-fsdp.yaml
4+
5+
model_class: llama.LlamaForCausalLM # Used to import the model from this class
6+
vocab_size: 128256
7+
hidden_size: 4096
8+
intermediate_size: 14336
9+
num_hidden_layers: 32
10+
num_attention_heads: 32
11+
num_key_value_heads: 8
12+
hidden_act: silu
13+
max_position_embeddings: 131072
14+
bos_token_id: 128000
15+
eos_token_id: 128001
16+
tokenizer_name: meta-llama/Meta-Llama-3.1-8B
17+
initializer_range: 0.02
18+
rms_norm_eps: 1.0e-05
19+
attention_dropout: false
20+
attention_bias: false
21+
flash_attention: true
22+
rope_theta: 500000.0
23+
rope_scaling:
24+
factor: 8.0
25+
low_freq_factor: 1.0
26+
high_freq_factor: 4.0
27+
original_context_len: 8192

0 commit comments

Comments
 (0)