1. 22 Oct, 2025 1 commit
    • thatPepe's avatar
      Issue/497 - Enhanced Test Framework (#520) · f5e6d729
      thatPepe authored
      * issue/497 - add dtype __eq__ and __hash__
      
      * issue/497 - simplified infinicore test functions
      
      * issue/497 - improved test framework
      
      greatly reduced the code required for specific operators;
      added strided tensor support;
      
      * issue/497 - add add interface to assist test
      
      * issue/497 - generalized test framework based on add
      
      * issue/497 - support non-contiguous tensors in result comparison
      
      * issue/497 - temporarily fixed strided tensor creation
      
      * issue/497 - rms norm interface
      
      * issue/497 - now requires test function definition
      
      * issue/497 - support mixed dtype
      
      * issue/497 - initial rms norm test
      
      * issue/497 - unified in place and out of place tests
      
      * issue/497 - renamed src/infinicore/op
      
      * issue/497 - reduced comments
      
      * issue/497 - attention
      
      * issue/497 - removed generic parameter mapping
      
      * issue/497 - temporary attention test
      
      * issue/497 - captitalize op name initial
      
      * issue/497 - add a script to run all op tests
      
      * issue/497 - fix comments
      
      * issue/497 - simplified infinicore tensor creation from torch
      
      * issue/497 - support tensor init modes
      
      * issue/497 - support tensor from/to files
      
      * issue/497 - adjust naming
      f5e6d729
  2. 21 Oct, 2025 1 commit
  3. 16 Oct, 2025 1 commit
  4. 11 Oct, 2025 2 commits
  5. 29 Sep, 2025 2 commits
  6. 25 Sep, 2025 1 commit
  7. 23 Sep, 2025 1 commit
  8. 18 Sep, 2025 2 commits
  9. 17 Sep, 2025 1 commit
  10. 16 Sep, 2025 2 commits
  11. 15 Sep, 2025 2 commits
  12. 10 Sep, 2025 1 commit
  13. 03 Sep, 2025 3 commits
  14. 02 Sep, 2025 1 commit
  15. 27 Aug, 2025 1 commit
  16. 22 Aug, 2025 1 commit
  17. 20 Aug, 2025 1 commit
  18. 14 Aug, 2025 1 commit
  19. 13 Aug, 2025 3 commits
  20. 12 Aug, 2025 1 commit
  21. 06 Aug, 2025 1 commit
  22. 14 Jul, 2025 1 commit
  23. 11 Jul, 2025 2 commits
  24. 10 Jul, 2025 1 commit
  25. 09 Jul, 2025 2 commits
  26. 08 Jul, 2025 1 commit
  27. 07 Jul, 2025 1 commit
  28. 04 Jul, 2025 1 commit
  29. 01 Jul, 2025 1 commit
    • 蒋帅宏(Shuaihong_Jiang)'s avatar
      issue/254: 添加算子在CPU和CUDA上对BF16的支持,并增加相应的测试代码 (#255) · f88d4ad8
      蒋帅宏(Shuaihong_Jiang) authored
      
      
      * issue/254: 添加算子在CPU和CUDA上对BF16的支持,并增加相应的测试代码
      
      * issue/254: 将修改后的算子格式化后重新提交
      
      * 修改与最新main的冲突
      
      * 解决冲突后rms_norm原本的精度过不了了,现在由
      {"atol": 5e-3, "rtol": 5e-3}更改为
      {"atol": 8e-3, "rtol": 8e-3}
      
      * rms_norm在debug模式下FP16的测试用例失败了(本地测试能通过,github上过不了),
      所以将容差增大了两倍进行测试
      
      * 将rms_normd的测试输入缩放0.5,将容差改回原始值来进行ci测试
      
      * issue/254: 1.使用CHECK_DTYPE宏来进行数据类型检验
      2.在test的utils.py中添加了设备对BF16支持的检验
      
      * issue/254: rms_norm测试fp16容差由
      torch.float16: {"atol": 1e-3, "rtol": 1e-3},
      改为torch.float16: {"atol": 2e-3, "rtol": 2e-3},
      并删除对输入0.5的放缩
      
      * issue/254: 在utils.py中debug方法和debug_all方法中
      添加了对BF16的特判
      
      * 修改支持BF16测试的设备类型检查方法
      
      * 修改支持BF16测试的设备检查
      
      * issue/254: reduce redundancy in rms_norm.py
      
      * issue/254: add back the missing comment in rms_norm.py
      
      * issue/254: add fp32 tolerance condition in causal_softmax.py
      
      ---------
      Co-authored-by: default avatarZimin Li <coollizimin@gmail.com>
      f88d4ad8