Unverified Commit c9745ee0 authored by Lianmin Zheng's avatar Lianmin Zheng Committed by GitHub
Browse files

Fix pandas dependency in CI (#3818)

parent 1a6e9757
...@@ -124,8 +124,8 @@ class ServerArgs: ...@@ -124,8 +124,8 @@ class ServerArgs:
speculative_draft_model_path: Optional[str] = None speculative_draft_model_path: Optional[str] = None
speculative_algorithm: Optional[str] = None speculative_algorithm: Optional[str] = None
speculative_num_steps: int = 5 speculative_num_steps: int = 5
speculative_num_draft_tokens: int = 64
speculative_eagle_topk: int = 8 speculative_eagle_topk: int = 8
speculative_num_draft_tokens: int = 64
# Double Sparsity # Double Sparsity
enable_double_sparsity: bool = False enable_double_sparsity: bool = False
...@@ -719,12 +719,6 @@ class ServerArgs: ...@@ -719,12 +719,6 @@ class ServerArgs:
help="The number of steps sampled from draft model in Speculative Decoding.", help="The number of steps sampled from draft model in Speculative Decoding.",
default=ServerArgs.speculative_num_steps, default=ServerArgs.speculative_num_steps,
) )
parser.add_argument(
"--speculative-num-draft-tokens",
type=int,
help="The number of token sampled from draft model in Speculative Decoding.",
default=ServerArgs.speculative_num_draft_tokens,
)
parser.add_argument( parser.add_argument(
"--speculative-eagle-topk", "--speculative-eagle-topk",
type=int, type=int,
...@@ -732,6 +726,12 @@ class ServerArgs: ...@@ -732,6 +726,12 @@ class ServerArgs:
choices=[1, 2, 4, 8], choices=[1, 2, 4, 8],
default=ServerArgs.speculative_eagle_topk, default=ServerArgs.speculative_eagle_topk,
) )
parser.add_argument(
"--speculative-num-draft-tokens",
type=int,
help="The number of token sampled from draft model in Speculative Decoding.",
default=ServerArgs.speculative_num_draft_tokens,
)
# Double Sparsity # Double Sparsity
parser.add_argument( parser.add_argument(
......
...@@ -20,7 +20,7 @@ pip install flashinfer_python==0.2.1.post2 --find-links ${FLASHINFER_REPO} --for ...@@ -20,7 +20,7 @@ pip install flashinfer_python==0.2.1.post2 --find-links ${FLASHINFER_REPO} --for
pip install torch_memory_saver --force-reinstall pip install torch_memory_saver --force-reinstall
pip install transformers==4.45.2 sentence_transformers accelerate peft pip install transformers==4.45.2 sentence_transformers accelerate peft pandas datasets
# For compling xgrammar kernels # For compling xgrammar kernels
pip install cuda-python nvidia-cuda-nvrtc-cu12 pip install cuda-python nvidia-cuda-nvrtc-cu12
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment