Commit 6b38250f authored by Jiezhong Qiu's avatar Jiezhong Qiu
Browse files

import from pytorch dir

parent 3501fa41
...@@ -2,7 +2,7 @@ import math ...@@ -2,7 +2,7 @@ import math
from torch import nn from torch import nn
import torch import torch
from moe_function import moe from .moe_function import moe
class MOELayer(nn.Module): class MOELayer(nn.Module):
......
...@@ -34,7 +34,7 @@ class PositionalEmbedding(nn.Module): ...@@ -34,7 +34,7 @@ class PositionalEmbedding(nn.Module):
return pos_emb[:,None,:] return pos_emb[:,None,:]
class CustomizedMoEPositionwiseFF(nn.Module): class CustomizedMoEPositionwiseFF(nn.Module):
def __init__(self, d_model, d_inner, dropout, pre_lnorm=False, top_k=2, num_expert=32): def __init__(self, d_model, d_inner, dropout, pre_lnorm=False, top_k=2, num_expert=64):
super(CustomizedMoEPositionwiseFF, self).__init__() super(CustomizedMoEPositionwiseFF, self).__init__()
print("CustomizedMoEPositionwiseFF num_expert=%d top_k=%d" % (num_expert, top_k)) print("CustomizedMoEPositionwiseFF num_expert=%d top_k=%d" % (num_expert, top_k))
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment