forked from mrq/bitsandbytes-rocm
debug
This commit is contained in:
parent
5d65817101
commit
d9b8789818
|
@ -545,6 +545,7 @@ def test_linear8bitlt_no_fp16_weights(threshold, memory_efficient_backward):
|
||||||
.to(torch.float16)
|
.to(torch.float16)
|
||||||
.to("cuda")
|
.to("cuda")
|
||||||
)
|
)
|
||||||
|
w1, w2 = mlp.fc1.weight.clone(), mlp.fc2.weight.clone()
|
||||||
|
|
||||||
for i in range(100):
|
for i in range(100):
|
||||||
b1 = torch.randn(16, 8, 32, device="cuda").half()
|
b1 = torch.randn(16, 8, 32, device="cuda").half()
|
||||||
|
@ -567,8 +568,15 @@ def test_linear8bitlt_no_fp16_weights(threshold, memory_efficient_backward):
|
||||||
assert o1.requires_grad
|
assert o1.requires_grad
|
||||||
grad_proj = torch.randn_like(o1)
|
grad_proj = torch.randn_like(o1)
|
||||||
|
|
||||||
|
mlp.zero_grad()
|
||||||
(o1 * grad_proj).sum().backward()
|
(o1 * grad_proj).sum().backward()
|
||||||
|
|
||||||
|
grad_ref = grad_proj.flatten(2) @ w2 @ w1
|
||||||
|
assert torch.allclose(b1.grad, grad_ref)
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue
Block a user