File tree Expand file tree Collapse file tree 1 file changed +27
-0
lines changed
torchprime/torch_xla_models/configs/model Expand file tree Collapse file tree 1 file changed +27
-0
lines changed Original file line number Diff line number Diff line change
1
+ defaults :
2
+ - _self_ # refers to this config file
3
+ - scaling : llama-fsdp # refers to scaling/llama-fsdp.yaml
4
+
5
+ model_class : llama.LlamaForCausalLM # Used to import the model from this class
6
+ vocab_size : 128256
7
+ hidden_size : 4096
8
+ intermediate_size : 14336
9
+ num_hidden_layers : 32
10
+ num_attention_heads : 32
11
+ num_key_value_heads : 8
12
+ hidden_act : silu
13
+ max_position_embeddings : 131072
14
+ bos_token_id : 128000
15
+ eos_token_id : 128001
16
+ tokenizer_name : meta-llama/Meta-Llama-3.1-8B
17
+ initializer_range : 0.02
18
+ rms_norm_eps : 1.0e-05
19
+ attention_dropout : false
20
+ attention_bias : false
21
+ flash_attention : true
22
+ rope_theta : 500000.0
23
+ rope_scaling :
24
+ factor : 8.0
25
+ low_freq_factor : 1.0
26
+ high_freq_factor : 4.0
27
+ original_context_len : 8192
You can’t perform that action at this time.
0 commit comments