Skip to content

Commit f5e51ba

Browse files
committed
Updated LoRA scaling, and learning rates.
1 parent 280e2a1 commit f5e51ba

File tree

2 files changed

+2
-2
lines changed

2 files changed

+2
-2
lines changed

src/thunder/config/adaptation/lora.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,6 +5,6 @@ lora_alpha: 16
55
# Hyperparameters
66
batch_size: 2
77
num_workers: 2
8-
lr: [1e-5, 1e-6, 1e-7]
8+
lr: [1e-3, 1e-4, 1e-5]
99
weight_decay: [0.0, 1e-3, 1e-4]
1010
epochs: 20

src/thunder/models/adapters.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -114,7 +114,7 @@ def __init__(
114114
std = torch.sqrt(torch.tensor(rank).float())
115115
self.A = torch.nn.Parameter(torch.randn(in_dim, rank) / std)
116116
self.B = torch.nn.Parameter(torch.zeros(rank, out_dim))
117-
self.alpha = alpha
117+
self.alpha = alpha / rank
118118

119119
def forward(self, src: torch.Tensor) -> torch.Tensor:
120120
"""

0 commit comments

Comments
 (0)