"git@developer.sourcefind.cn:renzhc/diffusers_dcu.git" did not exist on "9e00b727adc58ea8c744e68bccce7855e23dd10a"
Unverified Commit 0b9f64d6 authored by Hongzhi (Steve), Chen's avatar Hongzhi (Steve), Chen Committed by GitHub
Browse files
parent 815b88a6
...@@ -56,7 +56,7 @@ def evaluate(g, pred): ...@@ -56,7 +56,7 @@ def evaluate(g, pred):
return val_acc, test_acc return val_acc, test_acc
def train(g, A_hat, X, model): def train(model, g, A_hat, X):
label = g.ndata["label"] label = g.ndata["label"]
train_mask = g.ndata["train_mask"] train_mask = g.ndata["train_mask"]
optimizer = Adam(model.parameters(), lr=1e-2, weight_decay=5e-4) optimizer = Adam(model.parameters(), lr=1e-2, weight_decay=5e-4)
...@@ -114,4 +114,4 @@ if __name__ == "__main__": ...@@ -114,4 +114,4 @@ if __name__ == "__main__":
model = APPNP(in_size, out_size).to(dev) model = APPNP(in_size, out_size).to(dev)
# Kick off training. # Kick off training.
train(g, A_hat, X, model) train(model, g, A_hat, X)
...@@ -32,7 +32,7 @@ def evaluate(g, pred): ...@@ -32,7 +32,7 @@ def evaluate(g, pred):
return val_acc, test_acc return val_acc, test_acc
def train(g, X_sgc, model): def train(model, g, X_sgc):
label = g.ndata["label"] label = g.ndata["label"]
train_mask = g.ndata["train_mask"] train_mask = g.ndata["train_mask"]
optimizer = Adam(model.parameters(), lr=2e-1, weight_decay=5e-6) optimizer = Adam(model.parameters(), lr=2e-1, weight_decay=5e-6)
...@@ -91,4 +91,4 @@ if __name__ == "__main__": ...@@ -91,4 +91,4 @@ if __name__ == "__main__":
model = nn.Linear(in_size, out_size).to(dev) model = nn.Linear(in_size, out_size).to(dev)
# Kick off training. # Kick off training.
train(g, X_sgc, model) train(model, g, X_sgc)
...@@ -57,7 +57,7 @@ def evaluate(g, pred): ...@@ -57,7 +57,7 @@ def evaluate(g, pred):
return val_acc, test_acc return val_acc, test_acc
def train(g, X_sign, model): def train(model, g, X_sign):
label = g.ndata["label"] label = g.ndata["label"]
train_mask = g.ndata["train_mask"] train_mask = g.ndata["train_mask"]
optimizer = Adam(model.parameters(), lr=3e-3) optimizer = Adam(model.parameters(), lr=3e-3)
...@@ -124,4 +124,4 @@ if __name__ == "__main__": ...@@ -124,4 +124,4 @@ if __name__ == "__main__":
model = SIGN(in_size, out_size, r).to(dev) model = SIGN(in_size, out_size, r).to(dev)
# Kick off training. # Kick off training.
train(g, X_sign, model) train(model, g, X_sign)
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment