update tokenizer load
This commit is contained in:
parent
45c0d12049
commit
e34d4e9371
@ -33,10 +33,9 @@ def init_model(args):
|
|||||||
apply_lora(model)
|
apply_lora(model)
|
||||||
load_lora(model, f'./{args.out_dir}/lora/{args.lora_name}_{args.dim}.pth')
|
load_lora(model, f'./{args.out_dir}/lora/{args.lora_name}_{args.dim}.pth')
|
||||||
else:
|
else:
|
||||||
model = AutoModelForCausalLM.from_pretrained(
|
transformers_model_path = './MiniMind2'
|
||||||
'./MiniMind2',
|
tokenizer = AutoTokenizer.from_pretrained(transformers_model_path)
|
||||||
trust_remote_code=True
|
model = AutoModelForCausalLM.from_pretrained(transformers_model_path, trust_remote_code=True)
|
||||||
)
|
|
||||||
print(f'MiniMind模型参数量: {sum(p.numel() for p in model.parameters() if p.requires_grad) / 1e6:.2f}M(illion)')
|
print(f'MiniMind模型参数量: {sum(p.numel() for p in model.parameters() if p.requires_grad) / 1e6:.2f}M(illion)')
|
||||||
return model.eval().to(args.device), tokenizer
|
return model.eval().to(args.device), tokenizer
|
||||||
|
|
||||||
|
Loading…
x
Reference in New Issue
Block a user