Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
gaoqiong
lm-evaluation-harness
Commits
953bfdd2
Commit
953bfdd2
authored
Nov 23, 2023
by
baberabb
Browse files
Update vllm compatibility in README.
parent
b570fce7
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
11 additions
and
11 deletions
+11
-11
README.md
README.md
+11
-11
No files found.
README.md
View file @
953bfdd2
...
...
@@ -138,17 +138,17 @@ Our library also supports the evaluation of models served via several commercial
A full accounting of the supported and planned libraries + APIs can be seen below:
| API or Inference Server | Implemented? |
`--model <xxx>`
name | Models supported: | Request Types: |
|-----------------------------|---------------------------------|----------------------------------------------------------------------------------|--------------------------------------|----------------------------------------------------------|
| OpenAI Completions | :heavy_check_mark: |
`openai`
,
`openai-completions`
,
`gooseai`
| up to
`code-davinci-002`
|
`generate_until`
,
`loglikelihood`
,
`loglikelihood_rolling`
|
| OpenAI ChatCompletions | :x: Not yet - needs testing! | N/A |
[
All ChatCompletions API models
](
https://platform.openai.com/docs/guides/gpt
)
|
`generate_until`
(no logprobs) |
| Anthropic | :heavy_check_mark: |
`anthropic`
|
[
Supported Anthropic Engines
](
https://docs.anthropic.com/claude/reference/selecting-a-model
)
|
`generate_until`
(no logprobs) |
| GooseAI | :heavy_check_mark: (not separately maintained) |
`openai`
,
`openai-completions`
,
`gooseai`
(same interface as OpenAI Completions) | |
`generate_until`
,
`loglikelihood`
,
`loglikelihood_rolling`
|
| Textsynth | Needs testing |
`textsynth`
| ??? |
`generate_until`
,
`loglikelihood`
,
`loglikelihood_rolling`
|
| Cohere | :hourglass: - blocked on Cohere API bug | N/A |
[
All `cohere.generate()` engines
](
https://docs.cohere.com/docs/models
)
|
`generate_until`
,
`loglikelihood`
,
`loglikelihood_rolling`
|
| GGML/
[
Llama.cpp
](
https://github.com/ggerganov/llama.cpp
)
(
via
[
llama-cpp-python
]
(https://github.com/abetlen/llama-cpp-python)) | :heavy_check_mark: |
`gguf`
,
`ggml`
| Llama-architecture models (Llama, Llama 2, Llemma, Mistral(?), Llama finetunes)
|
`generate_until`
,
`loglikelihood`
,
`loglikelihood_rolling`
|
| vLLM | :
x: Not yet - needs help!
|
N/A
|
All HF models |
`generate_until`
(no logprobs)
|
| Your inference server here! | ... | ... | ... | ... | | ... |
| API or Inference Server | Implemented? |
`--model <xxx>`
name | Models supported:
| Request Types: |
|-----------------------------|---------------------------------|----------------------------------------------------------------------------------|--------------------------------------
---------------------------------------------------------
|----------------------------------------------------------|
| OpenAI Completions | :heavy_check_mark: |
`openai`
,
`openai-completions`
,
`gooseai`
| up to
`code-davinci-002`
|
`generate_until`
,
`loglikelihood`
,
`loglikelihood_rolling`
|
| OpenAI ChatCompletions | :x: Not yet - needs testing! | N/A |
[
All ChatCompletions API models
](
https://platform.openai.com/docs/guides/gpt
)
|
`generate_until`
(no logprobs) |
| Anthropic | :heavy_check_mark: |
`anthropic`
|
[
Supported Anthropic Engines
](
https://docs.anthropic.com/claude/reference/selecting-a-model
)
|
`generate_until`
(no logprobs) |
| GooseAI | :heavy_check_mark: (not separately maintained) |
`openai`
,
`openai-completions`
,
`gooseai`
(same interface as OpenAI Completions) |
|
`generate_until`
,
`loglikelihood`
,
`loglikelihood_rolling`
|
| Textsynth | Needs testing |
`textsynth`
| ???
|
`generate_until`
,
`loglikelihood`
,
`loglikelihood_rolling`
|
| Cohere | :hourglass: - blocked on Cohere API bug | N/A |
[
All `cohere.generate()` engines
](
https://docs.cohere.com/docs/models
)
|
`generate_until`
,
`loglikelihood`
,
`loglikelihood_rolling`
|
| GGML/
[
Llama.cpp
](
https://github.com/ggerganov/llama.cpp
)
(
via
[
llama-cpp-python
]
(https://github.com/abetlen/llama-cpp-python)) | :heavy_check_mark: |
`gguf`
,
`ggml`
| Llama-architecture models (Llama, Llama 2, Llemma, Mistral(?), Llama finetunes) |
`generate_until`
,
`loglikelihood`
,
`loglikelihood_rolling`
|
| vLLM | :
heavy_check_mark:
|
`vllm`
|
[
Most HF Causal Language Models
](
https://docs.vllm.ai/en/latest/models/supported_models.html
)
|
`generate_until`
,
`loglikelihood`
,
`loglikelihood_rolling`
|
| Your inference server here! | ... | ... | ...
| ... | | ... |
It is on our roadmap to create task variants designed to enable models which do not serve logprobs/loglikelihoods to be compared with generation performance of open-source models.
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment