Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
dc01cf9c
"git@developer.sourcefind.cn:chenpangpang/transformers.git" did not exist on "e10fb9cbe6117deda7aa7944276256a2954d3fd7"
Unverified
Commit
dc01cf9c
authored
Jan 12, 2024
by
Joao Gante
Committed by
GitHub
Jan 12, 2024
Browse files
Docs: add model paths (#28475)
parent
d0264988
Changes
5
Hide whitespace changes
Inline
Side-by-side
Showing
5 changed files
with
10 additions
and
10 deletions
+10
-10
src/transformers/models/deprecated/open_llama/modeling_open_llama.py
...rmers/models/deprecated/open_llama/modeling_open_llama.py
+2
-2
src/transformers/models/idefics/modeling_idefics.py
src/transformers/models/idefics/modeling_idefics.py
+2
-2
src/transformers/models/llama/modeling_llama.py
src/transformers/models/llama/modeling_llama.py
+2
-2
src/transformers/models/mistral/modeling_mistral.py
src/transformers/models/mistral/modeling_mistral.py
+2
-2
src/transformers/models/mixtral/modeling_mixtral.py
src/transformers/models/mixtral/modeling_mixtral.py
+2
-2
No files found.
src/transformers/models/deprecated/open_llama/modeling_open_llama.py
View file @
dc01cf9c
...
@@ -730,8 +730,8 @@ class OpenLlamaForCausalLM(OpenLlamaPreTrainedModel):
...
@@ -730,8 +730,8 @@ class OpenLlamaForCausalLM(OpenLlamaPreTrainedModel):
```python
```python
>>> from transformers import AutoTokenizer, OpenLlamaForCausalLM
>>> from transformers import AutoTokenizer, OpenLlamaForCausalLM
>>> model = OpenLlamaForCausalLM.from_pretrained(
PATH_TO_CONVERTED_WEIGHTS
)
>>> model = OpenLlamaForCausalLM.from_pretrained(
"openlm-research/open_llama_7b"
)
>>> tokenizer = AutoTokenizer.from_pretrained(
PATH_TO_CONVERTED_TOKENIZER
)
>>> tokenizer = AutoTokenizer.from_pretrained(
"openlm-research/open_llama_7b"
)
>>> prompt = "Hey, are you conscious? Can you talk to me?"
>>> prompt = "Hey, are you conscious? Can you talk to me?"
>>> inputs = tokenizer(prompt, return_tensors="pt")
>>> inputs = tokenizer(prompt, return_tensors="pt")
...
...
src/transformers/models/idefics/modeling_idefics.py
View file @
dc01cf9c
...
@@ -1492,8 +1492,8 @@ class IdeficsForVisionText2Text(IdeficsPreTrainedModel):
...
@@ -1492,8 +1492,8 @@ class IdeficsForVisionText2Text(IdeficsPreTrainedModel):
```python
```python
>>> from transformers import AutoTokenizer, IdeficsForVisionText2Text
>>> from transformers import AutoTokenizer, IdeficsForVisionText2Text
>>> model = IdeficsForVisionText2Text.from_pretrained(
PATH_TO_CONVERTED_WEIGHTS
)
>>> model = IdeficsForVisionText2Text.from_pretrained(
"HuggingFaceM4/idefics-9b"
)
>>> tokenizer = AutoTokenizer.from_pretrained(
PATH_TO_CONVERTED_TOKENIZER
)
>>> tokenizer = AutoTokenizer.from_pretrained(
"HuggingFaceM4/idefics-9b"
)
>>> prompt = "Hey, are you consciours? Can you talk to me?"
>>> prompt = "Hey, are you consciours? Can you talk to me?"
>>> inputs = tokenizer(prompt, return_tensors="pt")
>>> inputs = tokenizer(prompt, return_tensors="pt")
...
...
src/transformers/models/llama/modeling_llama.py
View file @
dc01cf9c
...
@@ -1162,8 +1162,8 @@ class LlamaForCausalLM(LlamaPreTrainedModel):
...
@@ -1162,8 +1162,8 @@ class LlamaForCausalLM(LlamaPreTrainedModel):
```python
```python
>>> from transformers import AutoTokenizer, LlamaForCausalLM
>>> from transformers import AutoTokenizer, LlamaForCausalLM
>>> model = LlamaForCausalLM.from_pretrained(
PATH_TO_CONVERTED_WEIGHTS
)
>>> model = LlamaForCausalLM.from_pretrained(
"meta-llama/Llama-2-7b-hf"
)
>>> tokenizer = AutoTokenizer.from_pretrained(
PATH_TO_CONVERTED_TOKENIZER
)
>>> tokenizer = AutoTokenizer.from_pretrained(
"meta-llama/Llama-2-7b-hf"
)
>>> prompt = "Hey, are you conscious? Can you talk to me?"
>>> prompt = "Hey, are you conscious? Can you talk to me?"
>>> inputs = tokenizer(prompt, return_tensors="pt")
>>> inputs = tokenizer(prompt, return_tensors="pt")
...
...
src/transformers/models/mistral/modeling_mistral.py
View file @
dc01cf9c
...
@@ -1132,8 +1132,8 @@ class MistralForCausalLM(MistralPreTrainedModel):
...
@@ -1132,8 +1132,8 @@ class MistralForCausalLM(MistralPreTrainedModel):
```python
```python
>>> from transformers import AutoTokenizer, MistralForCausalLM
>>> from transformers import AutoTokenizer, MistralForCausalLM
>>> model = MistralForCausalLM.from_pretrained(
PATH_TO_CONVERTED_WEIGHTS
)
>>> model = MistralForCausalLM.from_pretrained(
"mistralai/Mistral-7B-v0.1"
)
>>> tokenizer = AutoTokenizer.from_pretrained(
PATH_TO_CONVERTED_TOKENIZER
)
>>> tokenizer = AutoTokenizer.from_pretrained(
"mistralai/Mistral-7B-v0.1"
)
>>> prompt = "Hey, are you conscious? Can you talk to me?"
>>> prompt = "Hey, are you conscious? Can you talk to me?"
>>> inputs = tokenizer(prompt, return_tensors="pt")
>>> inputs = tokenizer(prompt, return_tensors="pt")
...
...
src/transformers/models/mixtral/modeling_mixtral.py
View file @
dc01cf9c
...
@@ -1291,8 +1291,8 @@ class MixtralForCausalLM(MixtralPreTrainedModel):
...
@@ -1291,8 +1291,8 @@ class MixtralForCausalLM(MixtralPreTrainedModel):
```python
```python
>>> from transformers import AutoTokenizer, MixtralForCausalLM
>>> from transformers import AutoTokenizer, MixtralForCausalLM
>>> model = MixtralForCausalLM.from_pretrained(
PATH_TO_CONVERTED_WEIGHTS
)
>>> model = MixtralForCausalLM.from_pretrained(
"mistralai/Mixtral-8x7B-v0.1"
)
>>> tokenizer = AutoTokenizer.from_pretrained(
PATH_TO_CONVERTED_TOKENIZER
)
>>> tokenizer = AutoTokenizer.from_pretrained(
"mistralai/Mixtral-8x7B-v0.1"
)
>>> prompt = "Hey, are you conscious? Can you talk to me?"
>>> prompt = "Hey, are you conscious? Can you talk to me?"
>>> inputs = tokenizer(prompt, return_tensors="pt")
>>> inputs = tokenizer(prompt, return_tensors="pt")
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment