"git@developer.sourcefind.cn:chenpangpang/transformers.git" did not exist on "e10fb9cbe6117deda7aa7944276256a2954d3fd7"
Unverified Commit dc01cf9c authored by Joao Gante's avatar Joao Gante Committed by GitHub
Browse files

Docs: add model paths (#28475)

parent d0264988
...@@ -730,8 +730,8 @@ class OpenLlamaForCausalLM(OpenLlamaPreTrainedModel): ...@@ -730,8 +730,8 @@ class OpenLlamaForCausalLM(OpenLlamaPreTrainedModel):
```python ```python
>>> from transformers import AutoTokenizer, OpenLlamaForCausalLM >>> from transformers import AutoTokenizer, OpenLlamaForCausalLM
>>> model = OpenLlamaForCausalLM.from_pretrained(PATH_TO_CONVERTED_WEIGHTS) >>> model = OpenLlamaForCausalLM.from_pretrained("openlm-research/open_llama_7b")
>>> tokenizer = AutoTokenizer.from_pretrained(PATH_TO_CONVERTED_TOKENIZER) >>> tokenizer = AutoTokenizer.from_pretrained("openlm-research/open_llama_7b")
>>> prompt = "Hey, are you conscious? Can you talk to me?" >>> prompt = "Hey, are you conscious? Can you talk to me?"
>>> inputs = tokenizer(prompt, return_tensors="pt") >>> inputs = tokenizer(prompt, return_tensors="pt")
......
...@@ -1492,8 +1492,8 @@ class IdeficsForVisionText2Text(IdeficsPreTrainedModel): ...@@ -1492,8 +1492,8 @@ class IdeficsForVisionText2Text(IdeficsPreTrainedModel):
```python ```python
>>> from transformers import AutoTokenizer, IdeficsForVisionText2Text >>> from transformers import AutoTokenizer, IdeficsForVisionText2Text
>>> model = IdeficsForVisionText2Text.from_pretrained(PATH_TO_CONVERTED_WEIGHTS) >>> model = IdeficsForVisionText2Text.from_pretrained("HuggingFaceM4/idefics-9b")
>>> tokenizer = AutoTokenizer.from_pretrained(PATH_TO_CONVERTED_TOKENIZER) >>> tokenizer = AutoTokenizer.from_pretrained("HuggingFaceM4/idefics-9b")
>>> prompt = "Hey, are you consciours? Can you talk to me?" >>> prompt = "Hey, are you consciours? Can you talk to me?"
>>> inputs = tokenizer(prompt, return_tensors="pt") >>> inputs = tokenizer(prompt, return_tensors="pt")
......
...@@ -1162,8 +1162,8 @@ class LlamaForCausalLM(LlamaPreTrainedModel): ...@@ -1162,8 +1162,8 @@ class LlamaForCausalLM(LlamaPreTrainedModel):
```python ```python
>>> from transformers import AutoTokenizer, LlamaForCausalLM >>> from transformers import AutoTokenizer, LlamaForCausalLM
>>> model = LlamaForCausalLM.from_pretrained(PATH_TO_CONVERTED_WEIGHTS) >>> model = LlamaForCausalLM.from_pretrained("meta-llama/Llama-2-7b-hf")
>>> tokenizer = AutoTokenizer.from_pretrained(PATH_TO_CONVERTED_TOKENIZER) >>> tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-2-7b-hf")
>>> prompt = "Hey, are you conscious? Can you talk to me?" >>> prompt = "Hey, are you conscious? Can you talk to me?"
>>> inputs = tokenizer(prompt, return_tensors="pt") >>> inputs = tokenizer(prompt, return_tensors="pt")
......
...@@ -1132,8 +1132,8 @@ class MistralForCausalLM(MistralPreTrainedModel): ...@@ -1132,8 +1132,8 @@ class MistralForCausalLM(MistralPreTrainedModel):
```python ```python
>>> from transformers import AutoTokenizer, MistralForCausalLM >>> from transformers import AutoTokenizer, MistralForCausalLM
>>> model = MistralForCausalLM.from_pretrained(PATH_TO_CONVERTED_WEIGHTS) >>> model = MistralForCausalLM.from_pretrained("mistralai/Mistral-7B-v0.1")
>>> tokenizer = AutoTokenizer.from_pretrained(PATH_TO_CONVERTED_TOKENIZER) >>> tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-v0.1")
>>> prompt = "Hey, are you conscious? Can you talk to me?" >>> prompt = "Hey, are you conscious? Can you talk to me?"
>>> inputs = tokenizer(prompt, return_tensors="pt") >>> inputs = tokenizer(prompt, return_tensors="pt")
......
...@@ -1291,8 +1291,8 @@ class MixtralForCausalLM(MixtralPreTrainedModel): ...@@ -1291,8 +1291,8 @@ class MixtralForCausalLM(MixtralPreTrainedModel):
```python ```python
>>> from transformers import AutoTokenizer, MixtralForCausalLM >>> from transformers import AutoTokenizer, MixtralForCausalLM
>>> model = MixtralForCausalLM.from_pretrained(PATH_TO_CONVERTED_WEIGHTS) >>> model = MixtralForCausalLM.from_pretrained("mistralai/Mixtral-8x7B-v0.1")
>>> tokenizer = AutoTokenizer.from_pretrained(PATH_TO_CONVERTED_TOKENIZER) >>> tokenizer = AutoTokenizer.from_pretrained("mistralai/Mixtral-8x7B-v0.1")
>>> prompt = "Hey, are you conscious? Can you talk to me?" >>> prompt = "Hey, are you conscious? Can you talk to me?"
>>> inputs = tokenizer(prompt, return_tensors="pt") >>> inputs = tokenizer(prompt, return_tensors="pt")
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment