Commit ee1f296e authored by JessicaOjo's avatar JessicaOjo
Browse files

remove chat completion -untested

parent b432d0e9
...@@ -85,7 +85,6 @@ class OpenaiCompletionsLM(TemplateLM): ...@@ -85,7 +85,6 @@ class OpenaiCompletionsLM(TemplateLM):
tokenizer: Optional[str] = None, tokenizer: Optional[str] = None,
tokenizer_backend: Literal["tiktoken", "huggingface"] = "tiktoken", tokenizer_backend: Literal["tiktoken", "huggingface"] = "tiktoken",
truncate: bool = False, truncate: bool = False,
chat: bool = False,
max_gen_toks: int = 256, max_gen_toks: int = 256,
batch_size: int = 1, batch_size: int = 1,
seed: int = 1234, seed: int = 1234,
...@@ -112,7 +111,6 @@ class OpenaiCompletionsLM(TemplateLM): ...@@ -112,7 +111,6 @@ class OpenaiCompletionsLM(TemplateLM):
self.base_url = base_url self.base_url = base_url
self.tokenizer_backend = tokenizer_backend self.tokenizer_backend = tokenizer_backend
self.truncate = truncate self.truncate = truncate
self.chat = chat
self._batch_size = int(batch_size) self._batch_size = int(batch_size)
self._max_gen_toks = max_gen_toks self._max_gen_toks = max_gen_toks
self._max_length = max_length self._max_length = max_length
...@@ -213,7 +211,6 @@ class OpenaiCompletionsLM(TemplateLM): ...@@ -213,7 +211,6 @@ class OpenaiCompletionsLM(TemplateLM):
response = oa_completion( response = oa_completion(
client=self.client, client=self.client,
model=self.model, model=self.model,
chat=self.chat,
prompt=inps, prompt=inps,
echo=True, echo=True,
max_tokens=0, max_tokens=0,
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment