Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
opencompass
Commits
8728287a
"...composable_kernel.git" did not exist on "cd167e492a8f85ec6f5965e50667e8a58d3aa3a1"
Unverified
Commit
8728287a
authored
Dec 28, 2023
by
bittersweet1999
Committed by
GitHub
Dec 28, 2023
Browse files
fix erro in configs (#750)
parent
81098722
Changes
10
Hide whitespace changes
Inline
Side-by-side
Showing
10 changed files
with
27 additions
and
27 deletions
+27
-27
configs/datasets/subjective_alignbench/alignbench_judgeby_judgelm.py
...asets/subjective_alignbench/alignbench_judgeby_judgelm.py
+1
-1
configs/models/judge_llm/auto_j/hf_autoj_bilingual_6b.py
configs/models/judge_llm/auto_j/hf_autoj_bilingual_6b.py
+2
-2
configs/models/judge_llm/auto_j/hf_autoj_eng_13b.py
configs/models/judge_llm/auto_j/hf_autoj_eng_13b.py
+5
-5
configs/models/judge_llm/auto_j/hf_autoj_eng_13b_4bit.py
configs/models/judge_llm/auto_j/hf_autoj_eng_13b_4bit.py
+6
-6
configs/models/judge_llm/auto_j/hf_autoj_scen_classifier.py
configs/models/judge_llm/auto_j/hf_autoj_scen_classifier.py
+2
-2
configs/models/judge_llm/judgelm/hf_judgelm_13b_v1.py
configs/models/judge_llm/judgelm/hf_judgelm_13b_v1.py
+2
-2
configs/models/judge_llm/judgelm/hf_judgelm_33b_v1.py
configs/models/judge_llm/judgelm/hf_judgelm_33b_v1.py
+3
-3
configs/models/judge_llm/judgelm/hf_judgelm_7b_v1.py
configs/models/judge_llm/judgelm/hf_judgelm_7b_v1.py
+2
-2
configs/models/judge_llm/pandalm/hf_alpaca_pandalm_7b_v1.py
configs/models/judge_llm/pandalm/hf_alpaca_pandalm_7b_v1.py
+2
-2
configs/models/judge_llm/pandalm/hf_pandalm_7b_v1.py
configs/models/judge_llm/pandalm/hf_pandalm_7b_v1.py
+2
-2
No files found.
configs/datasets/subjective_alignbench/alignbench_judgeby_judgelm.py
View file @
8728287a
...
@@ -28,7 +28,7 @@ for _name in subjective_all_sets:
...
@@ -28,7 +28,7 @@ for _name in subjective_all_sets:
]),
]),
),
),
retriever
=
dict
(
type
=
ZeroRetriever
),
retriever
=
dict
(
type
=
ZeroRetriever
),
inferencer
=
dict
(
type
=
GenInferencer
,
max_out_len
=
2048
),
inferencer
=
dict
(
type
=
GenInferencer
,
max_out_len
=
2048
),
)
)
subjective_eval_cfg
=
dict
(
subjective_eval_cfg
=
dict
(
...
...
configs/models/judge_llm/auto_j/hf_autoj_bilingual_6b.py
View file @
8728287a
...
@@ -16,8 +16,8 @@ models = [dict(
...
@@ -16,8 +16,8 @@ models = [dict(
truncation_side
=
'left'
,
truncation_side
=
'left'
,
trust_remote_code
=
True
,
trust_remote_code
=
True
,
use_fast
=
False
,),
use_fast
=
False
,),
max_out_len
=
512
,
max_out_len
=
1024
,
max_seq_len
=
2048
,
max_seq_len
=
4096
,
batch_size
=
8
,
batch_size
=
8
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
),
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
),
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
...
...
configs/models/judge_llm/auto_j/hf_autoj_eng_13b.py
View file @
8728287a
...
@@ -3,15 +3,15 @@ from opencompass.models import HuggingFaceCausalLM
...
@@ -3,15 +3,15 @@ from opencompass.models import HuggingFaceCausalLM
models
=
[
dict
(
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
type
=
HuggingFaceCausalLM
,
abbr
=
'autoj-13b
-GPTQ-4bits
'
,
abbr
=
'autoj-13b'
,
path
=
"GAIR/autoj-13b
-GPTQ-4bits
"
,
path
=
"GAIR/autoj-13b"
,
tokenizer_path
=
'GAIR/autoj-13b
-GPTQ-4bits
'
,
tokenizer_path
=
'GAIR/autoj-13b'
,
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
truncation_side
=
'left'
,
trust_remote_code
=
True
,
trust_remote_code
=
True
,
use_fast
=
False
,),
use_fast
=
False
,),
max_out_len
=
512
,
max_out_len
=
1024
,
max_seq_len
=
2048
,
max_seq_len
=
4096
,
batch_size
=
8
,
batch_size
=
8
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
),
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
),
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
...
...
configs/models/judge_llm/auto_j/hf_autoj_eng_13b_4bit.py
View file @
8728287a
...
@@ -8,16 +8,16 @@ https://huggingface.co/GAIR/autoj-13b-GPTQ-4bits
...
@@ -8,16 +8,16 @@ https://huggingface.co/GAIR/autoj-13b-GPTQ-4bits
models
=
[
dict
(
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
type
=
HuggingFaceCausalLM
,
abbr
=
'autoj-13b'
,
abbr
=
'autoj-13b
-GPTQ-4bits
'
,
path
=
"GAIR/autoj-13b"
,
path
=
"GAIR/autoj-13b
-GPTQ-4bits
"
,
tokenizer_path
=
'GAIR/autoj-13b'
,
tokenizer_path
=
'GAIR/autoj-13b
-GPTQ-4bits
'
,
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
truncation_side
=
'left'
,
trust_remote_code
=
True
,
trust_remote_code
=
True
,
use_fast
=
False
,),
use_fast
=
False
,),
max_out_len
=
512
,
max_out_len
=
1024
,
max_seq_len
=
2048
,
max_seq_len
=
4096
,
batch_size
=
8
,
batch_size
=
8
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
),
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
),
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
)]
)]
\ No newline at end of file
configs/models/judge_llm/auto_j/hf_autoj_scen_classifier.py
View file @
8728287a
...
@@ -10,8 +10,8 @@ models = [dict(
...
@@ -10,8 +10,8 @@ models = [dict(
truncation_side
=
'left'
,
truncation_side
=
'left'
,
trust_remote_code
=
True
,
trust_remote_code
=
True
,
use_fast
=
False
,),
use_fast
=
False
,),
max_out_len
=
512
,
max_out_len
=
1024
,
max_seq_len
=
2048
,
max_seq_len
=
4096
,
batch_size
=
8
,
batch_size
=
8
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
),
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
),
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
...
...
configs/models/judge_llm/judgelm/hf_judgelm_13b_v1.py
View file @
8728287a
...
@@ -10,8 +10,8 @@ models = [dict(
...
@@ -10,8 +10,8 @@ models = [dict(
truncation_side
=
'left'
,
truncation_side
=
'left'
,
trust_remote_code
=
True
,
trust_remote_code
=
True
,
use_fast
=
False
,),
use_fast
=
False
,),
max_out_len
=
512
,
max_out_len
=
1024
,
max_seq_len
=
2048
,
max_seq_len
=
4096
,
batch_size
=
8
,
batch_size
=
8
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
),
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
),
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
...
...
configs/models/judge_llm/judgelm/hf_judgelm_33b_v1.py
View file @
8728287a
...
@@ -10,9 +10,9 @@ models = [dict(
...
@@ -10,9 +10,9 @@ models = [dict(
truncation_side
=
'left'
,
truncation_side
=
'left'
,
trust_remote_code
=
True
,
trust_remote_code
=
True
,
use_fast
=
False
,),
use_fast
=
False
,),
max_out_len
=
512
,
max_out_len
=
1024
,
max_seq_len
=
2048
,
max_seq_len
=
4096
,
batch_size
=
8
,
batch_size
=
8
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
),
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
),
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
run_cfg
=
dict
(
num_gpus
=
4
,
num_procs
=
1
),
)]
)]
\ No newline at end of file
configs/models/judge_llm/judgelm/hf_judgelm_7b_v1.py
View file @
8728287a
...
@@ -10,8 +10,8 @@ models = [dict(
...
@@ -10,8 +10,8 @@ models = [dict(
truncation_side
=
'left'
,
truncation_side
=
'left'
,
trust_remote_code
=
True
,
trust_remote_code
=
True
,
use_fast
=
False
,),
use_fast
=
False
,),
max_out_len
=
512
,
max_out_len
=
1024
,
max_seq_len
=
2048
,
max_seq_len
=
4096
,
batch_size
=
8
,
batch_size
=
8
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
),
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
),
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
...
...
configs/models/judge_llm/pandalm/hf_alpaca_pandalm_7b_v1.py
View file @
8728287a
...
@@ -10,8 +10,8 @@ models = [dict(
...
@@ -10,8 +10,8 @@ models = [dict(
truncation_side
=
'left'
,
truncation_side
=
'left'
,
trust_remote_code
=
True
,
trust_remote_code
=
True
,
use_fast
=
False
,),
use_fast
=
False
,),
max_out_len
=
512
,
max_out_len
=
1024
,
max_seq_len
=
2048
,
max_seq_len
=
4096
,
batch_size
=
8
,
batch_size
=
8
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
),
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
),
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
...
...
configs/models/judge_llm/pandalm/hf_pandalm_7b_v1.py
View file @
8728287a
...
@@ -10,8 +10,8 @@ models = [dict(
...
@@ -10,8 +10,8 @@ models = [dict(
truncation_side
=
'left'
,
truncation_side
=
'left'
,
trust_remote_code
=
True
,
trust_remote_code
=
True
,
use_fast
=
False
,),
use_fast
=
False
,),
max_out_len
=
512
,
max_out_len
=
1024
,
max_seq_len
=
2048
,
max_seq_len
=
4096
,
batch_size
=
8
,
batch_size
=
8
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
),
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
),
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment