|
@ -255,8 +255,9 @@ if __name__ == "__main__": |
|
|
"mayaeary/pygmalion-6b-4bit-128g", local_files_only=True) |
|
|
"mayaeary/pygmalion-6b-4bit-128g", local_files_only=True) |
|
|
|
|
|
|
|
|
elif args.model_name == 'pygmalion-6b-gptq-4bit': |
|
|
elif args.model_name == 'pygmalion-6b-gptq-4bit': |
|
|
model = AutoModelForCausalLM.from_pretrained( |
|
|
model = load_quantized("pygmalion-6b-gptq-4bit", 4, 128, device).to(device) |
|
|
"OccamRazor/pygmalion-6b-gptq-4bit", local_files_only=True, from_pt=True).to(device) |
|
|
# model = AutoModelForCausalLM.from_pretrained( |
|
|
|
|
|
# "OccamRazor/pygmalion-6b-gptq-4bit", local_files_only=True, from_pt=True).to(device) |
|
|
tokenizer = AutoTokenizer.from_pretrained( |
|
|
tokenizer = AutoTokenizer.from_pretrained( |
|
|
"OccamRazor/pygmalion-6b-gptq-4bit", local_files_only=True) |
|
|
"OccamRazor/pygmalion-6b-gptq-4bit", local_files_only=True) |
|
|
|
|
|
|
|
|