AmberLJC commited on
Commit
d9e1a5d
·
verified ·
1 Parent(s): 6016155

Upload todo.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. todo.md +22 -0
todo.md ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Gradient Clipping Experiment
2
+
3
+ ## Objective
4
+ Demonstrate how gradient clipping stabilizes training by preventing sudden large weight updates caused by rare, high-loss data points.
5
+
6
+ ## Task Breakdown
7
+
8
+ - [ ] Step 1: Implement simple PyTorch model (Embedding + Linear)
9
+ - [ ] Step 2: Create imbalanced synthetic dataset (990 'A', 10 'B' targets)
10
+ - [ ] Step 3: Training loop WITHOUT gradient clipping - record metrics
11
+ - [ ] Step 4: Training loop WITH gradient clipping (threshold=1.0) - record metrics
12
+ - [ ] Step 5: Generate comparison plots
13
+ - [ ] Step 6: Write summary report with findings
14
+
15
+ ## Key Metrics to Track
16
+ 1. Training loss per step
17
+ 2. L2 norm of gradients (before clipping)
18
+ 3. L2 norm of model weights
19
+
20
+ ## Expected Outcome
21
+ - Without clipping: Spiky gradient norms when encountering rare 'B' samples
22
+ - With clipping: Bounded gradient norms, more stable training