Update app.py
Browse files
app.py
CHANGED
|
@@ -85,7 +85,7 @@ class CustomTrainerCallback(TrainerCallback):
|
|
| 85 |
ckpt_name = f"checkpoint-{step}"
|
| 86 |
ckpt_path = os.path.join(args.output_dir, ckpt_name)
|
| 87 |
|
| 88 |
-
job.add_log(f"System:
|
| 89 |
|
| 90 |
def _upload_bg():
|
| 91 |
try:
|
|
@@ -94,9 +94,9 @@ class CustomTrainerCallback(TrainerCallback):
|
|
| 94 |
path_in_repo=".",
|
| 95 |
repo_id=self.repo_id,
|
| 96 |
token=self.hf_token,
|
| 97 |
-
commit_message=f"Live
|
| 98 |
)
|
| 99 |
-
job.add_log(f"Cloud: Synced
|
| 100 |
except:
|
| 101 |
pass
|
| 102 |
|
|
@@ -221,7 +221,7 @@ def background_train_task(job_id, hf_token, model_name, new_repo_name,
|
|
| 221 |
optim="adamw_torch_fused" if torch.cuda.is_available() else "adamw_torch",
|
| 222 |
logging_steps=1,
|
| 223 |
save_strategy="steps",
|
| 224 |
-
save_steps=
|
| 225 |
save_total_limit=1,
|
| 226 |
report_to="none",
|
| 227 |
fp16=True if torch.cuda.is_available() else False,
|
|
@@ -343,7 +343,7 @@ def load_from_url(request: gr.Request):
|
|
| 343 |
pass
|
| 344 |
return gr.update(selected="launch_tab"), ""
|
| 345 |
|
| 346 |
-
with gr.Blocks(title="Nucleus Enterprise") as demo:
|
| 347 |
with gr.Column():
|
| 348 |
gr.Markdown("# ⚛️ NUCLEUS ENTERPRISE")
|
| 349 |
gr.Markdown("Autonomous LLM Foundry | V7.0 Scratch Edition")
|
|
|
|
| 85 |
ckpt_name = f"checkpoint-{step}"
|
| 86 |
ckpt_path = os.path.join(args.output_dir, ckpt_name)
|
| 87 |
|
| 88 |
+
job.add_log(f"System: 100-Step Snapshot saved ({ckpt_name})")
|
| 89 |
|
| 90 |
def _upload_bg():
|
| 91 |
try:
|
|
|
|
| 94 |
path_in_repo=".",
|
| 95 |
repo_id=self.repo_id,
|
| 96 |
token=self.hf_token,
|
| 97 |
+
commit_message=f"Live Checkpoint Step {step}"
|
| 98 |
)
|
| 99 |
+
job.add_log(f"Cloud: Synced Checkpoint {step} to Root")
|
| 100 |
except:
|
| 101 |
pass
|
| 102 |
|
|
|
|
| 221 |
optim="adamw_torch_fused" if torch.cuda.is_available() else "adamw_torch",
|
| 222 |
logging_steps=1,
|
| 223 |
save_strategy="steps",
|
| 224 |
+
save_steps=100,
|
| 225 |
save_total_limit=1,
|
| 226 |
report_to="none",
|
| 227 |
fp16=True if torch.cuda.is_available() else False,
|
|
|
|
| 343 |
pass
|
| 344 |
return gr.update(selected="launch_tab"), ""
|
| 345 |
|
| 346 |
+
with gr.Blocks(title="Nucleus Enterprise", theme=gr.themes.Base()) as demo:
|
| 347 |
with gr.Column():
|
| 348 |
gr.Markdown("# ⚛️ NUCLEUS ENTERPRISE")
|
| 349 |
gr.Markdown("Autonomous LLM Foundry | V7.0 Scratch Edition")
|