- 19 Jul, 2024 1 commit
-
-
Daniël de Kok authored
Deepseek V2 is a MoE model from Deepseek. Relevant variations compared to other models: - Grouped top-K in expert selection. - mscale in yarn is calculated using the `mscale` and `mscale_all_dim` configuration options. - `mscale_all_dim` is also used in scaling attention softmax. - Permuting of the query/key representations before applying rotary embeddings. - Some projections cannot be sharded (`q_a_proj`, `kv_a_proj_with_mqa`). So, we need weight loads that supports quantized weights. To this end `{Weights,WeightLoader}.get_weight` was added. - The query/key head dimensionality differs from that of the value, so we need to pad during attention. - Heads with size 192, needs an extension to our paged attention fork and we need to ensure that the KV cache is allocated with the correct size. - Shared experts.
-
- 25 Jun, 2024 1 commit
-
-
Daniël de Kok authored
* Add pytest release marker Annotate a test with `@pytest.mark.release` and it only gets run with `pytest integration-tests --release`. * Mark many models as `release` to speed up CI
-
- 27 May, 2024 1 commit
-
-
Daniël de Kok authored
-
- 21 Feb, 2024 1 commit
-
-
OlivierDehaene authored
-
- 02 Jun, 2023 1 commit
-
-
OlivierDehaene authored
Close #288
-
- 31 May, 2023 1 commit
-
-
OlivierDehaene authored
-
- 26 May, 2023 1 commit
-
-
OlivierDehaene authored
Co-authored-by:Joel Lamy-Poirier <joel.lamy-poirier@servicenow.com>
-
- 16 May, 2023 1 commit
-
-
OlivierDehaene authored
-
- 15 May, 2023 1 commit
-
-
OlivierDehaene authored
-