@@ -165,7 +165,7 @@ def guessed(model: 'LazyModel') -> 'Params':
165165 n_mult = 256 ,
166166 n_head = n_embd // 128 ,
167167 n_head_kv = n_embd // 128 ,
168- f_norm_eps = 1e-5 ,
168+ rms_norm_eps = 1e-5 ,
169169 n_layer = next (i for i in itertools .count () if f"model.layers.{ i } .self_attn.q_proj.weight" not in model ),
170170 )
171171
@@ -203,7 +203,7 @@ def loadHFTransformerJson(model: 'LazyModel', config_path: Path) -> 'Params':
203203 )
204204
205205 # LLaMA v2 70B params.json
206- # {"dim": 8192, "multiple_of": 4096, "ffn_dim_multiplier": 1.3, "n_heads": 64, "n_kv_heads": 8,
206+ # {"dim": 8192, "multiple_of": 4096, "ffn_dim_multiplier": 1.3, "n_heads": 64, "n_kv_heads": 8,
207207 # "n_layers": 80, "norm_eps": 1e-05, "vocab_size": -1}
208208 @staticmethod
209209 def loadOriginalParamsJson (model : 'LazyModel' , config_path : Path ) -> 'Params' :
@@ -230,8 +230,8 @@ def loadOriginalParamsJson(model: 'LazyModel', config_path: Path) -> 'Params':
230230 n_head = n_head ,
231231 n_head_kv = n_head_kv ,
232232 ffn_hidden_size = ffn_hidden_size ,
233- bos_token_id = bos_token_id ,
234- eos_token_id = eos_token_id ,
233+ bos_token_id = bos_token_id ,
234+ eos_token_id = eos_token_id ,
235235 )
236236
237237 @staticmethod
@@ -278,7 +278,7 @@ def __init__(self, fname_tokenizer: Path, params_vocab_size: int, fname_added_to
278278 def sentencepiece_tokens (self ) -> Iterable [Tuple [bytes , float ]]:
279279 tokenizer = self .sentencepiece_tokenizer
280280 for i in range (self .params_vocab_size ):
281- text : bytes
281+ text : bytes
282282 if i < tokenizer .vocab_size ():
283283 if tokenizer .is_unknown (i ):
284284 text = " \u2047 " .encode ("utf-8" )
@@ -1086,7 +1086,7 @@ def write_file_header(self, params: Params, file_type: NEFileType) -> None:
10861086 self .fout .write (struct .pack ("f" , params .rope_theta ))
10871087 self .fout .write (struct .pack ("f" , params .rope_scale ))
10881088
1089- # TODO, bos_token_id = 0 in https://huggingface.co/decapoda-research/llama-7b-hf/blob/main/config.json
1089+ # TODO, bos_token_id = 0 in https://huggingface.co/decapoda-research/llama-7b-hf/blob/main/config.json
10901090 # but bos_token_id = 1 in llama.cpp
10911091 self .fout .write (struct .pack ("i" , params .bos_token_id ))
10921092 self .fout .write (struct .pack ("i" , params .eos_token_id ))
@@ -1108,10 +1108,9 @@ def write_vocab(self, vocab: Vocab) -> None:
11081108
11091109 @staticmethod
11101110 def write_vocab_only (fname_out : Path , vocab : Vocab ) -> None :
1111+ params = Params (n_vocab = vocab .vocab_size , n_embd = 0 , n_mult = 0 , n_head = 1 , n_layer = 0 )
11111112 of = OutputFile (fname_out )
1112- params = Params (n_vocab = vocab .vocab_size , n_embd = 0 , n_mult = 0 , n_head = 1 , n_layer = 0 , file_type = NEFileType .AllF32 )
1113- of = OutputFile (fname_out )
1114- of .write_file_header (params )
1113+ of .write_file_header (params , file_type = NEFileType .AllF32 )
11151114 of .write_vocab (vocab )
11161115 of .fout .close ()
11171116
0 commit comments