mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2024-10-01 01:26:03 -04:00
Fix ExLlamaV2 loaders using unnecessary "bits" metadata
This commit is contained in:
parent
624faa1438
commit
db5f6cd1d8
@ -89,7 +89,8 @@ def get_model_metadata(model):
|
|||||||
if metadata['rope_scaling']['type'] == 'linear':
|
if metadata['rope_scaling']['type'] == 'linear':
|
||||||
model_settings['compress_pos_emb'] = metadata['rope_scaling']['factor']
|
model_settings['compress_pos_emb'] = metadata['rope_scaling']['factor']
|
||||||
|
|
||||||
if 'quantization_config' in metadata:
|
# Read GPTQ metadata for old GPTQ loaders
|
||||||
|
if 'quantization_config' in metadata and metadata['quantization_config'].get('quant_method', '') != 'exl2':
|
||||||
if 'bits' in metadata['quantization_config']:
|
if 'bits' in metadata['quantization_config']:
|
||||||
model_settings['wbits'] = metadata['quantization_config']['bits']
|
model_settings['wbits'] = metadata['quantization_config']['bits']
|
||||||
if 'group_size' in metadata['quantization_config']:
|
if 'group_size' in metadata['quantization_config']:
|
||||||
|
Loading…
Reference in New Issue
Block a user