joint-0.32-2.5 / config.yaml
joseph-tennyson's picture
Upload folder using huggingface_hub
35dac19 verified
config: /root/in-context-learning-GLM/src/glm_configs/joint_0.32_2.5.yaml
model:
family: gpt2
n_dims: 10
n_embd: 256
n_head: 8
n_layer: 12
n_positions: 40
out_dir: /root/in-context-learning-GLM/models/joint-0.32-2.5
test_run: false
training:
batch_size: 256
curriculum:
dims:
end: 10
inc: 0
interval: 10000
start: 10
points:
end: 40
inc: 0
interval: 10000
start: 40
data: gaussian
keep_every_steps: 4000
learning_rate: 0.00025
num_tasks: null
num_training_examples: null
resume_id: joint-0.32-2.5
save_every_steps: 4000
task: GLM
task_kwargs:
function_type:
- poisson
- neg_binomial
r: 2.5
scaling: 0.32
train_steps: 20000
wandb:
entity: in-context
log_every_steps: 10
name: joint-0.32-2.5
notes: ICL GLM training
project: in-context-training