"git@developer.sourcefind.cn:OpenDAS/mmcv.git" did not exist on "643009e4458109cb88ba5e669eec61a5e54c83be"
Commit 6a826c41 authored by justheuristic's avatar justheuristic
Browse files

pre-cast

parent d9b87898
...@@ -538,14 +538,11 @@ def test_linear8bitlt_no_fp16_weights(threshold, memory_efficient_backward): ...@@ -538,14 +538,11 @@ def test_linear8bitlt_no_fp16_weights(threshold, memory_efficient_backward):
assert mlp.fc1.weight.device.type == "cuda" assert mlp.fc1.weight.device.type == "cuda"
assert mlp.fc2.weight.device.type == "cuda" assert mlp.fc2.weight.device.type == "cuda"
mlp = ( mlp = MLP8bit(
MLP8bit(
32, 64, threshold=threshold, has_fp16_weights=False, memory_efficient_backward=memory_efficient_backward 32, 64, threshold=threshold, has_fp16_weights=False, memory_efficient_backward=memory_efficient_backward
) )
.to(torch.float16)
.to("cuda")
)
w1, w2 = mlp.fc1.weight.clone(), mlp.fc2.weight.clone() w1, w2 = mlp.fc1.weight.clone(), mlp.fc2.weight.clone()
mlp = mlp.cuda().half()
for i in range(100): for i in range(100):
b1 = torch.randn(16, 8, 32, device="cuda").half() b1 = torch.randn(16, 8, 32, device="cuda").half()
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment