"tests/test_structures/test_bbox/test_box3d.py" did not exist on "82cd4892fb8ed7a45e77e59241beeb8b531aada4"
- 15 Dec, 2025 1 commit
-
-
Parth Sareen authored
-
- 13 Dec, 2025 2 commits
-
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
- 12 Dec, 2025 2 commits
-
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
- 11 Dec, 2025 1 commit
-
-
Jeffrey Morgan authored
-
- 09 Dec, 2025 2 commits
-
-
nicole pardal authored
-
Jeffrey Morgan authored
-
- 08 Dec, 2025 1 commit
-
-
Michael Yang authored
change to a flatter directory structure and group the options with the function update models to call rope in one place
-
- 02 Dec, 2025 1 commit
-
-
Patrick Devine authored
This change: * fixes rope scaling in the mistral converter * updates ministral to include llama4 scaling * includes a new ministral parser for parsing reasoning and tool calling --------- Co-authored-by:jmorganca <jmorganca@gmail.com>
-
- 20 Nov, 2025 1 commit
-
-
Michael Yang authored
the check for mla omits v3 and r1 which should not return unsupported. instead check the tokenizer for compatibility
-
- 19 Nov, 2025 3 commits
-
-
Patrick Devine authored
-
nicole pardal authored
-
Michael Yang authored
-
- 18 Nov, 2025 1 commit
-
-
Grace authored
* Add mla for flash attention * Revert to using chunks
-
- 13 Nov, 2025 1 commit
-
-
Michael Yang authored
* use slice/chunks * bert * llama4 * gemma3n * gptoss * mistral3 * qwen3vl * qwen25vl * deepseek2 * remove unused ops
-
- 06 Nov, 2025 1 commit
-
-
Daniel Hiltgen authored
-
- 03 Nov, 2025 1 commit
-
-
Michael Yang authored
-
- 30 Oct, 2025 2 commits
-
-
Michael Yang authored
* ml(ggml): mrope * interleave mrope
-
Michael Yang authored
this change fixes images with an alpha channel by overlaying the image onto a white background
-
- 29 Oct, 2025 1 commit
-
-
Michael Yang authored
-
- 28 Oct, 2025 2 commits
-
-
Michael Yang authored
-
Michael Yang authored
-
- 18 Oct, 2025 1 commit
-
-
Daniel Hiltgen authored
Co-authored-by:Michael Yang <git@mxy.ng>
-
- 13 Oct, 2025 1 commit
-
-
Michael Yang authored
deepseek's qwen3 distill uses a different rope scheme so support both
-
- 09 Oct, 2025 2 commits
-
-
shengxinjing authored
-
shengxinjing authored
-
- 03 Oct, 2025 1 commit
-
-
Grace authored
-
- 24 Sep, 2025 1 commit
-
-
Grace authored
* init deepseek model file * temp removal of flash attention implementation * shapes and proper, can make a pass * query, key, value have good cosine similarity, but the max diff is a bit high * Attention block is working! ** with eager for now, have not added the mask line * Attention block is working! ** with eager for now, have not added the mask line * working MoE at around 0.95 cosine sim * added cosine similarity function * Starting end to end structure * Trying (and failing) to get rope to work, going to test full thing on tater * running on tater36... just not the right outputs * we have the right values for rope... but its still not working? * chnage Extrapolation Factor to 1 * removed adding residuals twice, removed normalization from shared expert, refactored Norms (Attention, MLP) to be outside the (Attention, MLP) blocks and in the Transformer block instead, add cache setLayer * Temporary modelfiles for cpu * change kpass intermediate step to kv, two layer outputs [0,1] look fine * this calls for 16 chicken nuggets * whoops * cleaning up code * delete stuff we dont need * getting rid of debug statements for llama cpp * working with long contexts * fix long context view error * reverting some changes I made for files that are not apart of pr * Added proper tokenizer for deeepseek3 * clean up model and go test * remove Modelfile * not passing the tests * whoops * how to pass the ci tests * resolving some of the comments * rename * linted and renamed deepseek3 -> deepseek2 * remove name go * addressed changes - main change was adopting qwen3 naming scheme * I cannot with linters * clean up logs * clean up logs --------- Co-authored-by:
Grace Guo <graceguo@Graces-MBP.localdomain> Co-authored-by:
Grace Guo <graceguo@Graces-MacBook-Pro.local> Co-authored-by:
graceguo <graceguo@tater36.localdomain>
-
- 23 Sep, 2025 2 commits
-
-
Michael Yang authored
-
Michael Yang authored
-
- 19 Sep, 2025 1 commit
-
-
Patrick Devine authored
* gemma: fix rope scaling for qat models * gofumpt yourself
-
- 18 Sep, 2025 1 commit
-
-
Michael Yang authored
* cleanup * use pooling.TypeNone * pooling test * qwen3 embed
-
- 17 Sep, 2025 1 commit
-
-
Michael Yang authored
* fix(llama): rope scale * spm llama * skip moe models * cleanup
-
- 16 Sep, 2025 2 commits
-
-
Michael Yang authored
* use ggml_*_split activations when possible * forward qkv
-
Michael Yang authored
* cleanup * use pooling.TypeNone * pooling test
-
- 15 Sep, 2025 2 commits
-
-
Michael Yang authored
* fix truncate * s/SentencePieceModel/SentencePiece/ * bert * wordpiece * refactor pooling * more tokenizers * normalize embeddings
-
Michael Yang authored
this cleans up the model interface slightly without too much impact in other areas
-
- 04 Sep, 2025 1 commit
-
-
Michael Yang authored
* ollama: add embeddings
-
- 29 Aug, 2025 1 commit
-
-
Daniel Hiltgen authored
* perf: build graph for next batch in parallel to keep GPU busy This refactors the main run loop of the ollama runner to perform the main GPU intensive tasks (Compute+Floats) in a go routine so we can prepare the next batch in parallel to reduce the amount of time the GPU stalls waiting for the next batch of work. * tests: tune integration tests for ollama engine This tunes the integration tests to focus more on models supported by the new engine.
-