Commit 77356fb2 authored by Baber's avatar Baber
Browse files

nit

parent a2bc6240
...@@ -13,8 +13,8 @@ Homepage: `https://github.com/NVIDIA/RULER` ...@@ -13,8 +13,8 @@ Homepage: `https://github.com/NVIDIA/RULER`
> [!NOTE] > [!NOTE]
> When using Ruler tasks, please note: > When using Ruler tasks, please note:
> 1. A tokenizer is required for data processing. The system will use the tokenizer from model_args, or fall back to the tokenizer associated with the pretrained model name. > 1. A tokenizer is required for data processing. The system will use the tokenizer from model_args, or fall back to the tokenizer associated with the pretrained model name.
> 2. The default maximum sequence length is 4096. For processing longer sequences, specify additional lengths using the metadata parameter: > 2. The default maximum sequence length is 4096. For calculating metrics of different max seq lengths, specify additional lengths using the metadata parameter:
> `--metadata=max_seq_length=4096,8192,16384,32768,65536,131072`. The metadata parameter can also be passed to the TaskManager. > `--metadata=max_seq_length=4096,8192,16384,32768,65536,131072`. The metadata parameter can also be passed to the TaskManager (metadata: dict[str, tuple]).
> 3. To prevent truncation of longer sequences, we recommend setting the max_length parameter in model_args: > 3. To prevent truncation of longer sequences, we recommend setting the max_length parameter in model_args:
> `--model_args=pretrained=...,max_length=32768` > `--model_args=pretrained=...,max_length=32768`
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment