update tokenizer load

This commit is contained in:
gongjy 2025-02-19 23:24:29 +08:00
parent 45c0d12049
commit e34d4e9371

View File

@ -33,10 +33,9 @@ def init_model(args):
apply_lora(model) apply_lora(model)
load_lora(model, f'./{args.out_dir}/lora/{args.lora_name}_{args.dim}.pth') load_lora(model, f'./{args.out_dir}/lora/{args.lora_name}_{args.dim}.pth')
else: else:
model = AutoModelForCausalLM.from_pretrained( transformers_model_path = './MiniMind2'
'./MiniMind2', tokenizer = AutoTokenizer.from_pretrained(transformers_model_path)
trust_remote_code=True model = AutoModelForCausalLM.from_pretrained(transformers_model_path, trust_remote_code=True)
)
print(f'MiniMind模型参数量: {sum(p.numel() for p in model.parameters() if p.requires_grad) / 1e6:.2f}M(illion)') print(f'MiniMind模型参数量: {sum(p.numel() for p in model.parameters() if p.requires_grad) / 1e6:.2f}M(illion)')
return model.eval().to(args.device), tokenizer return model.eval().to(args.device), tokenizer