Add lit tests; shows initial bfloat16 support for tt.dot#253
Add lit tests; shows initial bfloat16 support for tt.dot#253abrown merged 2 commits intokernelize-ai:mainfrom
bfloat16 support for tt.dot#253Conversation
The "expand to target type" logic added in kernelize-ai#231 works for `bfloat16` types, not just `float16`. This change adds lit tests to show the general form of the lowering our `convert-triton-cpu-to-llvm` does when run.
|
I will note that performance for |
bfloat16 supportbfloat16 support for tt.dot
alexbaden
left a comment
There was a problem hiding this comment.
These look fine, though once #254 lands I imagine you would have to update the fmul and fadd bits. Might be easier to just check for promotion and assume the promoted value is used (best yet canonicalize in the RUN command so an unused fp extension gets dropped) .
|
I tried to additionally add: But I think they have to be sequential? I guess I can figure that out in a follow-on PR. |
alexbaden
left a comment
There was a problem hiding this comment.
I'm fine with this for now with the caveats that long term (1) test_core.py::test_dot is preferred for correctness and (2) we should strive to make lit tests as focused as possible (e.g. only test promotion to fp32 during the llvm lowering).
The "expand to target type" logic added in #231 works for
bfloat16types, not justfloat16. This change adds lit tests to show the general form of the lowering ourconvert-triton-cpu-to-llvmdoes when run.