aadarsh99 commited on
Commit
1589eac
·
1 Parent(s): 9269be3

added new models

Browse files
fine_tuned_sam2_batched_18000.torch CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1db4f02a40e147f81a05f87fac10f06423c179d4d26d6e8b836ea49873b7ea27
3
  size 898098003
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fbfe698ea07ac6b8e394fd978bccdb1fb35aeb9daeafed272be3f8183133744
3
  size 898098003
fine_tuned_sam2_batched_plm_18000.torch CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:72f1b0161950f378b20403763860b7551a4b2280cbce36da55a553a6197da35d
3
  size 7661168597
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2478ccd1c8eec086560befcbd1b62f34ffff0e2ec8f601dc4860ac6b78d43ee1
3
  size 7661168597
lora_plm_adapter_18000/adapter_config.json CHANGED
@@ -25,12 +25,12 @@
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
28
- "gate_proj",
29
- "up_proj",
30
  "down_proj",
31
- "q_proj",
32
  "o_proj",
 
33
  "k_proj",
 
34
  "v_proj"
35
  ],
36
  "target_parameters": null,
 
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
 
 
28
  "down_proj",
29
+ "gate_proj",
30
  "o_proj",
31
+ "q_proj",
32
  "k_proj",
33
+ "up_proj",
34
  "v_proj"
35
  ],
36
  "target_parameters": null,
lora_plm_adapter_18000/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8d457f025fc5a609514dde7b4e2ad0adc99b02b275b05a695d87c7d0ababf669
3
  size 148712776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f10dc4870b85d3a3d243cab6e5e1124ec8b572c6b2c0b33bdb6b62962e54451a
3
  size 148712776