| from transformers import AutoModelForCausalLM |
| import torch |
| from modelscope import ( |
| snapshot_download, AutoModelForCausalLM, AutoTokenizer, GenerationConfig |
| ) |
| import torch |
| model_id = 'qwen/Qwen-VL-Chat' |
| revision = 'v1.0.3' |
|
|
| model_dir = snapshot_download(model_id, revision=revision) |
| model = AutoModelForCausalLM.from_pretrained(model_dir, device_map="auto", trust_remote_code=True, fp16=True).eval() |
|
|
| state_dict = model.state_dict() |
| save_dict = {} |
| for k,v in state_dict.items(): |
| if 'visual' in k: |
| if 'transformer.visual.proj' not in k: |
| save_dict[k.replace('transformer.visual.', '')] = v |
| torch.save(save_dict, './qwen_clip/pytorch_model.bin') |