DTee8 commited on
Commit
dc7baae
·
verified ·
1 Parent(s): 01ca456

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +61 -1
README.md CHANGED
@@ -4,4 +4,64 @@ language:
4
  - en
5
  base_model:
6
  - microsoft/Phi-4-multimodal-instruct
7
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  - en
5
  base_model:
6
  - microsoft/Phi-4-multimodal-instruct
7
+ ---
8
+
9
+ ##Galactus
10
+ This model is a fine-tuned version of microsoft/Phi-4-multimodal-instruct on the Galaxy's Last Exam Benchmark.
11
+
12
+ Galaxy's Last Exam Benchmark: 72%
13
+
14
+ ##Model description
15
+ Galactus is a state-of-the-art (SOTA) multimodal language model that outperforms all OpenAI and Gemini models on the Galaxy's Last Exam Benchmark. This benchmark features challenging tasks that push the boundaries of metaphysical competence—for instance, determining how many times two lines intersect or simulating the effect of adding three minutes to an analog clock.
16
+ The model accepts image input along with text prompts and has been specifically optimized to tackle complex visual reasoning tasks.
17
+
18
+ ##Intended uses & limitations
19
+ This model is intended for handling complex visual reasoning tasks that require metaphysical competence.
20
+
21
+ ##Training and evaluation data
22
+ The model was exclusively trained on the Galaxy's Last Exam Benchmark.
23
+
24
+ ##Training procedure
25
+ The model was trained using LoRA adapters focused on the vision components of the base model.
26
+
27
+ ##Prompt format
28
+ This model uses the following image prompt format:
29
+ Copy<|image_1|> + user text
30
+
31
+ ##Training hyperparameters
32
+ The following hyperparameters were used during training:
33
+
34
+ num_train_epochs: specified in args
35
+ per_device_train_batch_size: specified in args
36
+ gradient_checkpointing: True
37
+ gradient_checkpointing_kwargs: {'use_reentrant': False}
38
+ gradient_accumulation_steps: specified in args
39
+ optim: 'adamw_torch'
40
+ adam_beta1: 0.9
41
+ adam_beta2: 0.95
42
+ adam_epsilon: 1e-7
43
+ learning_rate: specified in args
44
+ weight_decay: 0.0 or as specified in args
45
+ save_strategy: 'steps'
46
+ save_steps: 10
47
+ eval_steps: 10 if eval_dataset else None
48
+ evaluation_strategy: 'steps' if eval_dataset else 'no'
49
+ load_best_model_at_end: True if eval_dataset else False
50
+ max_grad_norm: 1.0
51
+ lr_scheduler_type: 'linear'
52
+ warmup_steps: 50
53
+ logging_steps: 10
54
+ save_total_limit: 2
55
+ save_only_model: True
56
+ dataloader_num_workers: 4
57
+ ddp_find_unused_parameters: True
58
+
59
+ ##Training results
60
+ The model achieved 72% performance on the Galaxy's Last Exam Benchmark.
61
+
62
+ ##Framework versions
63
+
64
+ Transformers 4.46.1
65
+ PyTorch 2.7.0.dev20250304+cu128
66
+ TorchVision 0.22.0.dev20250304+cu128
67
+ Tokenizers 0.20.3