TanVir17Niloy commited on
Commit
8d07b7e
·
verified ·
1 Parent(s): ea4cdf0

Upload BLIP HyperModulator v1.0.0 - 2025-08-29 02:09

Browse files
Files changed (2) hide show
  1. README.md +3 -3
  2. model_metadata.json +1 -1
README.md CHANGED
@@ -32,8 +32,8 @@ from blip_hyper_modulator import BlipHyperModulator
32
  from sentence_transformers import SentenceTransformer
33
 
34
  # Load base BLIP model
35
- blip_model = BlipForConditionalGeneration.from_pretrained("Salesforce/blip-image-captioning-base")
36
- processor = BlipProcessor.from_pretrained("Salesforce/blip-image-captioning-base")
37
 
38
  # Load hypermodulator
39
  hypermodulator = BlipHyperModulator.from_pretrained("path/to/model", blip_model)
@@ -51,7 +51,7 @@ lora_weights = hypermodulator.generate_lora_weights(
51
 
52
  ## Model Details
53
 
54
- - **Base Model**: Salesforce/blip-image-captioning-base
55
  - **Created**: 2025-08-29
56
  - **Version**: 1.0.0
57
  - **Framework**: PyTorch
 
32
  from sentence_transformers import SentenceTransformer
33
 
34
  # Load base BLIP model
35
+ blip_model = BlipForConditionalGeneration.from_pretrained("Salesforce/blip-image-captioning-large")
36
+ processor = BlipProcessor.from_pretrained("Salesforce/blip-image-captioning-large")
37
 
38
  # Load hypermodulator
39
  hypermodulator = BlipHyperModulator.from_pretrained("path/to/model", blip_model)
 
51
 
52
  ## Model Details
53
 
54
+ - **Base Model**: Salesforce/blip-image-captioning-large
55
  - **Created**: 2025-08-29
56
  - **Version**: 1.0.0
57
  - **Framework**: PyTorch
model_metadata.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "model_name": "BLIP HyperModulator",
3
  "version": "1.0.0",
4
- "created_date": "2025-08-29T01:24:43.252168",
5
  "description": "Enhanced hypernetwork for generating LoRA adapters for BLIP image captioning models",
6
  "features": {
7
  "cross_attention": true,
 
1
  {
2
  "model_name": "BLIP HyperModulator",
3
  "version": "1.0.0",
4
+ "created_date": "2025-08-29T02:09:41.769218",
5
  "description": "Enhanced hypernetwork for generating LoRA adapters for BLIP image captioning models",
6
  "features": {
7
  "cross_attention": true,