{ "experiments": { "test_exp_001": { "id": "test_exp_001", "name": "Test Experiment", "description": "Debug test", "created_at": "2025-07-20T14:01:48.871089", "status": "running", "metrics": [ { "timestamp": "2025-07-20T14:01:48.871096", "step": 25, "metrics": { "loss": 1.165, "accuracy": 0.75, "learning_rate": 3.5e-06 } } ], "parameters": {}, "artifacts": [], "logs": [] }, "exp_20250720_130853": { "id": "exp_20250720_130853", "name": "petite-elle-l-aime-3", "description": "SmolLM3 fine-tuning experiment", "created_at": "2025-07-20T11:20:01.780908", "status": "running", "metrics": [ { "timestamp": "2025-07-20T11:20:01.780908", "step": 25, "metrics": { "loss": 1.1659, "grad_norm": 10.3125, "learning_rate": 7e-08, "num_tokens": 1642080.0, "mean_token_accuracy": 0.75923578992486, "epoch": 0.004851130919895701 } }, { "timestamp": "2025-07-20T11:26:39.042155", "step": 50, "metrics": { "loss": 1.165, "grad_norm": 10.75, "learning_rate": 1.4291666666666667e-07, "num_tokens": 3324682.0, "mean_token_accuracy": 0.7577659255266189, "epoch": 0.009702261839791402 } }, { "timestamp": "2025-07-20T11:33:16.203045", "step": 75, "metrics": { "loss": 1.1639, "grad_norm": 10.6875, "learning_rate": 2.1583333333333334e-07, "num_tokens": 4987941.0, "mean_token_accuracy": 0.7581205774843692, "epoch": 0.014553392759687101 } }, { "timestamp": "2025-07-20T11:39:53.453917", "step": 100, "metrics": { "loss": 1.1528, "grad_norm": 10.75, "learning_rate": 2.8875e-07, "num_tokens": 6630190.0, "mean_token_accuracy": 0.7614579878747463, "epoch": 0.019404523679582803 } } ], "parameters": { "model_name": "HuggingFaceTB/SmolLM3-3B", "max_seq_length": 12288, "use_flash_attention": true, "use_gradient_checkpointing": false, "batch_size": 8, "gradient_accumulation_steps": 16, "learning_rate": 3.5e-06, "weight_decay": 0.01, "warmup_steps": 1200, "max_iters": 18000, "eval_interval": 1000, "log_interval": 25, "save_interval": 2000, "optimizer": "adamw_torch", "beta1": 0.9, "beta2": 0.999, "eps": 1e-08, "scheduler": "cosine", "min_lr": 3.5e-07, "fp16": false, "bf16": true, "ddp_backend": "nccl", "ddp_find_unused_parameters": false, "save_steps": 2000, "eval_steps": 1000, "logging_steps": 25, "save_total_limit": 5, "eval_strategy": "steps", "metric_for_best_model": "eval_loss", "greater_is_better": false, "load_best_model_at_end": true, "data_dir": null, "train_file": null, "validation_file": null, "test_file": null, "use_chat_template": true, "chat_template_kwargs": { "add_generation_prompt": true, "no_think_system_message": true }, "enable_tracking": true, "trackio_url": "https://tonic-test-trackio-test.hf.space", "trackio_token": null, "log_artifacts": true, "log_metrics": true, "log_config": true, "experiment_name": "petite-elle-l-aime-3", "dataset_name": "legmlai/openhermes-fr", "dataset_split": "train", "input_field": "prompt", "target_field": "accepted_completion", "filter_bad_entries": true, "bad_entry_field": "bad_entry", "packing": false, "max_prompt_length": 12288, "max_completion_length": 8192, "truncation": true, "dataloader_num_workers": 10, "dataloader_pin_memory": true, "dataloader_prefetch_factor": 3, "max_grad_norm": 1.0, "group_by_length": true }, "artifacts": [], "logs": [] }, "exp_20250720_134319": { "id": "exp_20250720_134319", "name": "petite-elle-l-aime-3-1", "description": "SmolLM3 fine-tuning experiment", "created_at": "2025-07-20T11:54:31.993219", "status": "running", "metrics": [ { "timestamp": "2025-07-20T11:54:31.993219", "step": 25, "metrics": { "loss": 1.166, "grad_norm": 10.375, "learning_rate": 7e-08, "num_tokens": 1642080.0, "mean_token_accuracy": 0.7590958896279335, "epoch": 0.004851130919895701 } }, { "timestamp": "2025-07-20T11:54:33.589487", "step": 25, "metrics": { "gpu_0_memory_allocated": 17.202261447906494, "gpu_0_memory_reserved": 75.474609375, "gpu_0_utilization": 0, "cpu_percent": 2.7, "memory_percent": 10.1 } } ], "parameters": { "model_name": "HuggingFaceTB/SmolLM3-3B", "max_seq_length": 12288, "use_flash_attention": true, "use_gradient_checkpointing": false, "batch_size": 8, "gradient_accumulation_steps": 16, "learning_rate": 3.5e-06, "weight_decay": 0.01, "warmup_steps": 1200, "max_iters": 18000, "eval_interval": 1000, "log_interval": 25, "save_interval": 2000, "optimizer": "adamw_torch", "beta1": 0.9, "beta2": 0.999, "eps": 1e-08, "scheduler": "cosine", "min_lr": 3.5e-07, "fp16": false, "bf16": true, "ddp_backend": "nccl", "ddp_find_unused_parameters": false, "save_steps": 2000, "eval_steps": 1000, "logging_steps": 25, "save_total_limit": 5, "eval_strategy": "steps", "metric_for_best_model": "eval_loss", "greater_is_better": false, "load_best_model_at_end": true, "data_dir": null, "train_file": null, "validation_file": null, "test_file": null, "use_chat_template": true, "chat_template_kwargs": { "add_generation_prompt": true, "no_think_system_message": true }, "enable_tracking": true, "trackio_url": "https://tonic-test-trackio-test.hf.space", "trackio_token": null, "log_artifacts": true, "log_metrics": true, "log_config": true, "experiment_name": "petite-elle-l-aime-3-1", "dataset_name": "legmlai/openhermes-fr", "dataset_split": "train", "input_field": "prompt", "target_field": "accepted_completion", "filter_bad_entries": true, "bad_entry_field": "bad_entry", "packing": false, "max_prompt_length": 12288, "max_completion_length": 8192, "truncation": true, "dataloader_num_workers": 10, "dataloader_pin_memory": true, "dataloader_prefetch_factor": 3, "max_grad_norm": 1.0, "group_by_length": true }, "artifacts": [], "logs": [] } }, "current_experiment": "exp_20250720_134319", "last_updated": "2025-07-20T14:05:18.615415" }