"...csrc/git@developer.sourcefind.cn:change/sglang.git" did not exist on "63c8d8d0627924633664781f5469bf5dd68a0228"
Unverified Commit 0d46b175 authored by Praateek Mahajan's avatar Praateek Mahajan Committed by GitHub
Browse files

Update Readme

Incorrect link for `Quick tour: Fine-tuning/usage scripts`
parent 71d597da
...@@ -19,7 +19,7 @@ These implementations have been tested on several datasets (see the example scri ...@@ -19,7 +19,7 @@ These implementations have been tested on several datasets (see the example scri
|-|-| |-|-|
| [Installation](#installation) | How to install the package | | [Installation](#installation) | How to install the package |
| [Quick tour: Usage](#quick-tour-usage) | Tokenizers & models usage: Bert and GPT-2 | | [Quick tour: Usage](#quick-tour-usage) | Tokenizers & models usage: Bert and GPT-2 |
| [Quick tour: Fine-tuning/usage scripts](#quick-tour-fine-tuningusage-scripts) | Using provided scripts: GLUE, SQuAD and Text generation | | [Quick tour: Fine-tuning/usage scripts](#quick-tour-of-the-fine-tuningusage-scripts) | Using provided scripts: GLUE, SQuAD and Text generation |
| [Migrating from pytorch-pretrained-bert to pytorch-transformers](#Migrating-from-pytorch-pretrained-bert-to-pytorch-transformers) | Migrating your code from pytorch-pretrained-bert to pytorch-transformers | | [Migrating from pytorch-pretrained-bert to pytorch-transformers](#Migrating-from-pytorch-pretrained-bert-to-pytorch-transformers) | Migrating your code from pytorch-pretrained-bert to pytorch-transformers |
| [Documentation](https://huggingface.co/pytorch-transformers/) | Full API documentation and more | | [Documentation](https://huggingface.co/pytorch-transformers/) | Full API documentation and more |
...@@ -118,7 +118,6 @@ tokenizer = tokenizer_class.from_pretrained(pretrained_weights) ...@@ -118,7 +118,6 @@ tokenizer = tokenizer_class.from_pretrained(pretrained_weights)
``` ```
## Quick tour of the fine-tuning/usage scripts ## Quick tour of the fine-tuning/usage scripts
The library comprises several example scripts with SOTA performances for NLU and NLG tasks: The library comprises several example scripts with SOTA performances for NLU and NLG tasks:
- `run_glue.py`: an example fine-tuning Bert, XLNet and XLM on nine different GLUE tasks (*sequence-level classification*) - `run_glue.py`: an example fine-tuning Bert, XLNet and XLM on nine different GLUE tasks (*sequence-level classification*)
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment