"next_docs/en/git@developer.sourcefind.cn:wangsen/mineru.git" did not exist on "ea8f7e0fbd8e8b907482b95b12edc8a4093a5c40"
Unverified Commit ee1bff06 authored by Harry Wang's avatar Harry Wang Committed by GitHub
Browse files

minor docs grammar fixes (#6889)

parent 8abd7f69
......@@ -128,7 +128,7 @@ The encoded versions have different lengths:
>>> len(encoded_sequence_a), len(encoded_sequence_b)
(8, 19)
Therefore, we can't be put then together in a same tensor as-is. The first sequence needs to be padded up to the length
Therefore, we can't put them together in the same tensor as-is. The first sequence needs to be padded up to the length
of the second one, or the second one needs to be truncated down to the length of the first one.
In the first case, the list of IDs will be extended by the padding indices. We can pass a list to the tokenizer and ask
......
......@@ -214,7 +214,7 @@ Using the model
Once your input has been preprocessed by the tokenizer, you can send it directly to the model. As we mentioned, it will
contain all the relevant information the model needs. If you're using a TensorFlow model, you can pass the
dictionary keys directly to tensor, for a PyTorch model, you need to unpack the dictionary by adding :obj:`**`.
dictionary keys directly to tensors, for a PyTorch model, you need to unpack the dictionary by adding :obj:`**`.
.. code-block::
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment