test_gpt3.py 6.59 KB
Newer Older
Leo Gao's avatar
Leo Gao committed
1
2
3
4
5
6
7
8
9
import lm_eval.models as models
import pytest
import os
import json
import openai
import mock
import pickle
import hashlib

Leo Gao's avatar
Leo Gao committed
10

11
12
13
14
def mock_completion(**kwargs):
    # Mock completion function
    # Loads from a cached+pickled response if it exists, otherwise it will actually try to ping
    os.makedirs("tests/testdata", exist_ok=True)
Fabrizio Milo's avatar
Fabrizio Milo committed
15
16
17
    hash = hashlib.sha256(
        json.dumps(kwargs, sort_keys=True).encode("utf-8")
    ).hexdigest()
Leo Gao's avatar
Leo Gao committed
18
19
20
    fname = f"tests/testdata/gpt3_test_{hash}.pkl"

    if os.path.exists(fname):
Fabrizio Milo's avatar
Fabrizio Milo committed
21
        with open(fname, "rb") as fh:
Leo Gao's avatar
Leo Gao committed
22
23
            return pickle.load(fh)
    ret = openai.Completion.create(**kwargs)
24
    ret.api_key = ""
Fabrizio Milo's avatar
Fabrizio Milo committed
25
    with open(fname, "wb") as fh:
Leo Gao's avatar
Leo Gao committed
26
27
28
29
        pickle.dump(ret, fh)
    return ret


30
@mock.patch("lm_eval.models.gpt3.oa_completion", new=mock_completion)
Leo Gao's avatar
Leo Gao committed
31
def test_gpt3():
Fabrizio Milo's avatar
Fabrizio Milo committed
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
    if "OPENAI_API_SECRET_KEY" not in os.environ:
        os.environ["OPENAI_API_SECRET_KEY"] = ""
    gpt3 = models.get_model("gpt3").create_from_arg_string("engine=ada")
    (
        (ll_dog, ig_dog),
        (ll_cat, ig_cat),
        (_, ll_max_0),
        (_, ll_max_1),
        (_, ll_max_2),
        *vals,
    ) = gpt3.loglikelihood(
        [
            ("The quick brown fox jumps over the lazy", " dog"),
            ("The quick brown fox jumps over the lazy", " cat"),
            ("The quick brown fox jumps over the lazy", ", lazy dog"),
            ("The quick brown fox jumps over the lazy", ", lazy fox"),
            (
                "The quick brown fox jumps over the lazy",
                ", lazy fox and they both fall to the ground",
            ),
            (
                """A mult""",
                """ilayer perceptron (MLP) is a class of feedforward artificial neural network (ANN)""",
            ),
            (
                """The term MLP is used ambiguously, sometimes loosely to any feedforward ANN, sometimes strictly to refer to networks composed of multiple layers of perceptrons""",
                """ (with threshold activation); see § Terminology""",
            ),
            (
                """Multilayer perceptrons are sometimes coll""",
                """oquially referred to as "vanilla" neural networks, especially when they have a single hidden layer.[1]""",
            ),
            (
                """An MLP consists of at least three layers of nodes: an input layer, a hidden layer and an output layer. Except for the input nodes, each node is a neuron that uses a nonlinear""",
                """ activation function.""",
            ),
            (
                """MLP utilizes a supervised""",
                """ learning technique called backpropagation for training.[2][3] Its multiple layers and non-linear activation distinguish MLP from a linear perceptron. It can distinguish data that is not linearly separable.[4]""",
            ),
            (
                """Recent work has demonstrated substantial gains on many NLP tasks and benchmarks by pre-training on a large corpus of text followed by fine-tuning on a specific task. While typically task-agnostic""",
                """ in architecture, this method still requires task-specific fine-tuning datasets of thousands or tens of thousands of examples. By contrast, humans can generally perform a new language task from only a few examples or from simple instructions - something which current NLP systems still largely struggle to do. Here we show that scaling up language models greatly improves task-agnostic, few-shot performance, sometimes even reaching competitiveness with prior state-of-the-art fine-tuning approaches. """,
            ),
            (
                """Specifically, we train GPT-3, an autoregressive language model with 175""",
                """ billion parameters, 10x more than any previous non-sparse language model, and test its performance in the few-shot setting. For all tasks, GPT-3 is applied without any gradient updates or fine-tuning, with tasks and few-shot demonstrations specified purely via text interaction with the model. GPT-3 achieves strong performance on many NLP datasets, including translation, question-answering, and cloze tasks, as well as several tasks that require on-the-fly reasoning or domain adaptation, such as unscrambling words, using a novel word in a sentence, or performing 3-digit arithmetic. At the same time, we also identify some datasets where GPT-3's few-shot learning still struggles, as well as some datasets where GPT-3 faces methodological issues related to training on large web corpora. Finally, we find that GPT-3 can generate samples of news articles which human evaluators have difficulty distinguishing from articles written by humans. We discuss broader societal impacts of this finding and of GPT-3 in general.""",
            ),
            (
                """A mult""",
                """ilayer perceptron (MLP) is a class of feedforward artificial neural network (ANN)""",
            ),
            ("""Hello""", """ World"""),
        ]
    )
Leo Gao's avatar
Leo Gao committed
87
88
89
90
91
92
93
94
95
96

    assert ll_dog > ll_cat
    assert not ig_cat

    assert ig_dog
    assert not ll_max_0
    assert not ll_max_1
    assert not ll_max_2

    # test empty context
Fabrizio Milo's avatar
Fabrizio Milo committed
97
    gpt3.loglikelihood([("", "test")])
Leo Gao's avatar
Leo Gao committed
98

Fabrizio Milo's avatar
Fabrizio Milo committed
99
100
101
    (gen,) = gpt3.greedy_until(
        [("The quick brown fox jumps over the lazy", [".", "\n"])]
    )
Leo Gao's avatar
Leo Gao committed
102

Fabrizio Milo's avatar
Fabrizio Milo committed
103
    assert gen == " dog"
Leo Gao's avatar
Leo Gao committed
104
105
106

    print([x[0] for x in vals])

107
    targets = [
Fabrizio Milo's avatar
Fabrizio Milo committed
108
109
110
111
112
113
114
115
116
        -34.848301606999996,
        -47.148329679999996,
        -45.44380149599999,
        -5.285246016,
        -133.97821690686004,
        -321.2616693239001,
        -658.0299524401041,
        -34.848301606999996,
        -7.525115,
117
    ]
Leo Gao's avatar
Leo Gao committed
118
119
120
121
122

    for (pred, _), tgt in zip(vals, targets):
        assert pred == pytest.approx(tgt, rel=1e-3)


123
@mock.patch("lm_eval.models.gpt3.oa_completion", new=mock_completion)
Leo Gao's avatar
Leo Gao committed
124
def test_gpt3_perplexity():
Fabrizio Milo's avatar
Fabrizio Milo committed
125
126
127
    if "OPENAI_API_SECRET_KEY" not in os.environ:
        os.environ["OPENAI_API_SECRET_KEY"] = ""
    gpt3 = models.get_model("gpt3").create_from_arg_string("engine=ada")
Leo Gao's avatar
Leo Gao committed
128
129
130
131
132
133
    test_string = "We study empirical scaling laws for language model performance on the cross-entropy loss."
    perplexity = gpt3.loglikelihood_rolling([(test_string,)])[0]
    tgt = -84.38819608
    assert perplexity == pytest.approx(tgt, rel=1e-3)

    # Hack: modify gpt3 to have shorter context length to induce rolling windows
Fabrizio Milo's avatar
Fabrizio Milo committed
134
135
136
    with mock.patch.object(
        models.gpt3.GPT3LM, "max_length", new_callable=mock.PropertyMock
    ) as mock_max_length:
137
        mock_max_length.return_value = 5
Fabrizio Milo's avatar
Fabrizio Milo committed
138
        gpt3 = models.get_model("gpt3").create_from_arg_string("engine=ada")
139
140
        perplexity = gpt3.loglikelihood_rolling([(test_string,)])[0]
    tgt = -101.81967209999999
Leo Gao's avatar
Leo Gao committed
141
    assert perplexity == pytest.approx(tgt, rel=1e-3)