@@ -153,8 +153,7 @@ def write_tensors(self):
153
153
# map tensor names
154
154
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
155
155
if new_name is None :
156
- logger .error (f"Can not map tensor { name !r} " )
157
- sys .exit ()
156
+ raise ValueError (f"Can not map tensor { name !r} " )
158
157
159
158
n_dims = len (data .shape )
160
159
data_dtype = data .dtype
@@ -486,8 +485,7 @@ def write_tensors(self):
486
485
# map tensor names
487
486
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
488
487
if new_name is None :
489
- logger .error (f"Can not map tensor { name !r} " )
490
- sys .exit ()
488
+ raise ValueError (f"Can not map tensor { name !r} " )
491
489
492
490
n_dims = len (data .shape )
493
491
data_dtype = data .dtype
@@ -570,8 +568,7 @@ def write_tensors(self):
570
568
else :
571
569
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
572
570
if new_name is None :
573
- logger .error (f"Can not map tensor { name !r} " )
574
- sys .exit ()
571
+ raise ValueError (f"Can not map tensor { name !r} " )
575
572
576
573
n_dims = len (data .shape )
577
574
data_dtype = data .dtype
@@ -614,8 +611,7 @@ def set_gguf_parameters(self):
614
611
elif "model_max_length" in self .hparams :
615
612
ctx_length = self .hparams ["model_max_length" ]
616
613
else :
617
- logger .error ("gguf: can not find ctx length parameter." )
618
- sys .exit ()
614
+ raise ValueError ("gguf: can not find ctx length parameter." )
619
615
620
616
self .gguf_writer .add_file_type (self .ftype )
621
617
self .gguf_writer .add_name (self .dir_model .name )
@@ -653,8 +649,7 @@ def write_tensors(self):
653
649
# map tensor names
654
650
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
655
651
if new_name is None :
656
- logger .error (f"Can not map tensor { name !r} " )
657
- sys .exit ()
652
+ raise ValueError (f"Can not map tensor { name !r} " )
658
653
659
654
n_dims = len (data .shape )
660
655
data_dtype = data .dtype
@@ -696,8 +691,7 @@ def set_gguf_parameters(self):
696
691
elif "model_max_length" in self .hparams :
697
692
ctx_length = self .hparams ["model_max_length" ]
698
693
else :
699
- logger .error ("gguf: can not find ctx length parameter." )
700
- sys .exit ()
694
+ raise ValueError ("gguf: can not find ctx length parameter." )
701
695
702
696
self .gguf_writer .add_name (self .dir_model .name )
703
697
self .gguf_writer .add_source_hf_repo (hf_repo )
@@ -751,8 +745,7 @@ def write_tensors(self):
751
745
# map tensor names
752
746
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
753
747
if new_name is None :
754
- logger .error (f"Can not map tensor { name !r} " )
755
- sys .exit ()
748
+ raise ValueError (f"Can not map tensor { name !r} " )
756
749
757
750
n_dims = len (data .shape )
758
751
data_dtype = data .dtype
@@ -853,8 +846,7 @@ def set_gguf_parameters(self):
853
846
elif "model_max_length" in self .hparams :
854
847
ctx_length = self .hparams ["model_max_length" ]
855
848
else :
856
- logger .error ("gguf: can not find ctx length parameter." )
857
- sys .exit ()
849
+ raise ValueError ("gguf: can not find ctx length parameter." )
858
850
859
851
self .gguf_writer .add_name (self .dir_model .name )
860
852
self .gguf_writer .add_source_hf_repo (hf_repo )
@@ -903,8 +895,7 @@ def write_tensors(self):
903
895
# map tensor names
904
896
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
905
897
if new_name is None :
906
- logger .error (f"Can not map tensor { name !r} " )
907
- sys .exit ()
898
+ raise ValueError (f"Can not map tensor { name !r} " )
908
899
909
900
n_dims = len (data .shape )
910
901
data_dtype = data .dtype
@@ -1008,8 +999,7 @@ def write_tensors(self):
1008
999
# map tensor names
1009
1000
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
1010
1001
if new_name is None :
1011
- logger .error (f"Can not map tensor { name !r} " )
1012
- sys .exit ()
1002
+ raise ValueError (f"Can not map tensor { name !r} " )
1013
1003
1014
1004
n_dims = len (data .shape )
1015
1005
data_dtype = data .dtype
@@ -1111,10 +1101,9 @@ def write_tensors(self):
1111
1101
data = data_torch .squeeze ().numpy ()
1112
1102
1113
1103
# map tensor names
1114
- new_name = tensor_map .get_name (name , try_suffixes = (".weight" ,))
1104
+ new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
1115
1105
if new_name is None :
1116
- logger .error (f"Can not map tensor { name !r} " )
1117
- sys .exit ()
1106
+ raise ValueError (f"Can not map tensor { name !r} " )
1118
1107
1119
1108
n_dims = len (data .shape )
1120
1109
data_dtype = data .dtype
@@ -1180,8 +1169,7 @@ def write_tensors(self):
1180
1169
data = data_torch .to (torch .float32 ).squeeze ().numpy ()
1181
1170
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
1182
1171
if new_name is None :
1183
- logger .error (f"Can not map tensor { name !r} " )
1184
- sys .exit ()
1172
+ raise ValueError (f"Can not map tensor { name !r} " )
1185
1173
n_dims = len (data .shape )
1186
1174
logger .info (f"{ new_name } , n_dims = { n_dims } , { old_dtype } --> { data .dtype } " )
1187
1175
self .gguf_writer .add_tensor (new_name , data )
@@ -1383,8 +1371,7 @@ def write_tensors(self):
1383
1371
1384
1372
new_name = tensor_map .get_name (merged_name , try_suffixes = (".weight" , ".bias" ))
1385
1373
if new_name is None :
1386
- logger .error (f"Can not map tensor { name !r} " )
1387
- sys .exit ()
1374
+ raise ValueError (f"Can not map tensor { name !r} " )
1388
1375
1389
1376
logger .info (f"{ new_name } , n_dims = { len (data .shape )} , shape = { data .shape } --> { data .dtype } " )
1390
1377
@@ -1394,8 +1381,7 @@ def write_tensors(self):
1394
1381
# map tensor names
1395
1382
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
1396
1383
if new_name is None :
1397
- logger .error (f"Can not map tensor { name !r} " )
1398
- sys .exit ()
1384
+ raise ValueError (f"Can not map tensor { name !r} " )
1399
1385
1400
1386
n_dims = len (data .shape )
1401
1387
data_dtype = data .dtype
@@ -1487,8 +1473,7 @@ def write_tensors(self):
1487
1473
1488
1474
new_name = tensor_map .get_name (merged_name , try_suffixes = (".weight" , ".bias" ))
1489
1475
if new_name is None :
1490
- logger .error (f"Can not map tensor { name !r} " )
1491
- sys .exit ()
1476
+ raise ValueError (f"Can not map tensor { name !r} " )
1492
1477
1493
1478
logger .info (f"{ new_name } , n_dims = { len (data .shape )} , shape = { data .shape } --> { data .dtype } " )
1494
1479
@@ -1498,8 +1483,7 @@ def write_tensors(self):
1498
1483
# map tensor names
1499
1484
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
1500
1485
if new_name is None :
1501
- logger .error (f"Can not map tensor { name !r} " )
1502
- sys .exit ()
1486
+ raise ValueError (f"Can not map tensor { name !r} " )
1503
1487
1504
1488
n_dims = len (data .shape )
1505
1489
data_dtype = data .dtype
@@ -1592,17 +1576,15 @@ def write_tensors(self):
1592
1576
# https://huggingface.co/databricks/dbrx-instruct/blob/main/model.safetensors.index.json#L15
1593
1577
new_name = tensor_map .get_name (name if not experts else name + ".weight" , try_suffixes = (".weight" ,))
1594
1578
if new_name is None :
1595
- logger .error (f"Can not map tensor { name !r} " )
1596
- sys .exit ()
1579
+ raise ValueError (f"Can not map tensor { name !r} " )
1597
1580
1598
1581
n_dims = len (data .shape )
1599
1582
data_dtype = data .dtype
1600
1583
1601
1584
# Most of the codebase that takes in 1D tensors only handles F32 tensors
1602
1585
# and most of the outputs tensors are F32.
1603
1586
if data_dtype != np .float32 and n_dims == 1 :
1604
- logger .error (f"Can not map tensor { name !r} : all 1D tensors must be F32" )
1605
- sys .exit ()
1587
+ raise ValueError (f"Can not map tensor { name !r} : all 1D tensors must be F32" )
1606
1588
1607
1589
# if f32 desired, convert any float16 to float32
1608
1590
if self .ftype == 0 and data_dtype == np .float16 :
@@ -1674,8 +1656,7 @@ def write_tensors(self):
1674
1656
# map tensor names
1675
1657
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
1676
1658
if new_name is None :
1677
- logger .error (f"Can not map tensor { name !r} " )
1678
- sys .exit ()
1659
+ raise ValueError (f"Can not map tensor { name !r} " )
1679
1660
1680
1661
n_dims = len (data .shape )
1681
1662
data_dtype = data .dtype
@@ -1758,8 +1739,7 @@ def write_tensors(self):
1758
1739
# map tensor names
1759
1740
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
1760
1741
if new_name is None :
1761
- logger .error (f"Can not map tensor { name !r} " )
1762
- sys .exit ()
1742
+ raise ValueError (f"Can not map tensor { name !r} " )
1763
1743
1764
1744
n_dims = len (data .shape )
1765
1745
data_dtype = data .dtype
@@ -1921,8 +1901,7 @@ def write_tensors(self):
1921
1901
# map tensor names
1922
1902
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
1923
1903
if new_name is None :
1924
- logger .error (f"Can not map tensor { name !r} " )
1925
- sys .exit ()
1904
+ raise ValueError (f"Can not map tensor { name !r} " )
1926
1905
1927
1906
n_dims = len (data .shape )
1928
1907
data_dtype = data .dtype
@@ -2019,8 +1998,7 @@ def write_tensors(self):
2019
1998
# map tensor names
2020
1999
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
2021
2000
if new_name is None :
2022
- logger .error (f"Can not map tensor { name !r} " )
2023
- sys .exit ()
2001
+ raise ValueError (f"Can not map tensor { name !r} " )
2024
2002
2025
2003
# shuffle for broadcasting of gqa in ggml_mul_mat
2026
2004
if new_name .endswith ("attn_q.weight" ):
@@ -2097,8 +2075,7 @@ def write_tensors(self):
2097
2075
# map tensor names
2098
2076
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
2099
2077
if new_name is None :
2100
- logger .error (f"Can not map tensor { name !r} " )
2101
- sys .exit ()
2078
+ raise ValueError (f"Can not map tensor { name !r} " )
2102
2079
2103
2080
n_dims = len (data .shape )
2104
2081
data_dtype = data .dtype
@@ -2243,8 +2220,7 @@ def post_write_tensors(self, tensor_map, name, data_torch):
2243
2220
# map tensor names
2244
2221
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
2245
2222
if new_name is None :
2246
- logger .error (f"Can not map tensor { name !r} " )
2247
- sys .exit ()
2223
+ raise ValueError (f"Can not map tensor { name !r} " )
2248
2224
2249
2225
n_dims = len (data .shape )
2250
2226
data_dtype = data .dtype
@@ -2371,8 +2347,7 @@ def write_tensors(self):
2371
2347
# map tensor names
2372
2348
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
2373
2349
if new_name is None :
2374
- logger .error (f"Can not map tensor { name !r} " )
2375
- sys .exit ()
2350
+ raise ValueError (f"Can not map tensor { name !r} " )
2376
2351
2377
2352
data = data_torch .squeeze ().numpy ()
2378
2353
n_dims = len (data .shape )
@@ -2481,8 +2456,7 @@ def write_tensors(self):
2481
2456
# map tensor names
2482
2457
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
2483
2458
if new_name is None :
2484
- logger .error (f"Can not map tensor { name !r} " )
2485
- sys .exit ()
2459
+ raise ValueError (f"Can not map tensor { name !r} " )
2486
2460
2487
2461
n_dims = len (data .shape )
2488
2462
data_dtype = data .dtype
@@ -2584,8 +2558,7 @@ def write_tensors(self):
2584
2558
# map tensor names
2585
2559
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
2586
2560
if new_name is None :
2587
- logger .error (f"Can not map tensor { name !r} " )
2588
- sys .exit ()
2561
+ raise ValueError (f"Can not map tensor { name !r} " )
2589
2562
2590
2563
if name .endswith (".A_log" ):
2591
2564
logger .debug ("A_log --> A ==> " + new_name )
0 commit comments