WCNegentropy commited on
Commit
08981d9
·
verified ·
1 Parent(s): 3a300bc

Remove nested directory: BitTransformerLM/bit_transformer/optimization.py

Browse files
BitTransformerLM/bit_transformer/optimization.py DELETED
@@ -1,37 +0,0 @@
1
- import torch
2
- import torch.nn as nn
3
- from torch.optim import AdamW
4
- from torch.optim.lr_scheduler import OneCycleLR
5
-
6
-
7
- def configure_optimizer(
8
- model: nn.Module,
9
- lr: float = 1e-3,
10
- weight_decay: float = 0.01,
11
- total_steps: int = 100
12
- ):
13
- """Return AdamW optimizer with OneCycleLR scheduler."""
14
- optimizer = AdamW(model.parameters(), lr=lr, weight_decay=weight_decay)
15
- scheduler = OneCycleLR(optimizer, max_lr=lr, total_steps=total_steps)
16
- return optimizer, scheduler
17
-
18
-
19
- def adjust_learning_rate(optimizer: torch.optim.Optimizer, factor: float) -> float:
20
- """Scale the learning rate of all param groups by ``factor``.
21
-
22
- Parameters
23
- ----------
24
- optimizer:
25
- The optimizer whose learning rate will be adjusted.
26
- factor:
27
- Multiplicative factor applied to the current learning rate.
28
-
29
- Returns
30
- -------
31
- float
32
- The updated learning rate of the first parameter group.
33
- """
34
- for param_group in optimizer.param_groups:
35
- param_group["lr"] *= factor
36
- return optimizer.param_groups[0]["lr"]
37
-