nod-ai / iree-amd-aie

IREE plugin repository for the AMD AIE accelerator
Apache License 2.0
69 stars 30 forks source link

[xrt-lite] Update cts test to use objectFifo with bf16. #890

Closed newling closed 3 days ago

newling commented 1 week ago

Motivation: let's test objectFifo as that is now the pipeline we're supporting long-term.

makslevental commented 1 week ago

It should be as robust as the codegen itself? I've tested it locally up to like 1024x1024. One thing to be careful about is stale build artifacts because of the labyrythine method used to generate the vmfb. So try clearing the build dir that contains xrt_lite_executables_c.h.

I'm trying to adapt/use it to make a standalone matmul benchmarking test.

No advice here but why doesn't iree-e2e-benchmark-module suffice?

newling commented 1 week ago

It should be as robust as the codegen itself? I've tested it locally up to like 1024x1024. One thing to be careful about is stale build artifacts because of the labyrythine method used to generate the vmfb. So try clearing the build dir that contain

I'm fairly sure it's not that, I give the mlir func a new name before runs. But I'll double check.

No advice here but why doesn't iree-e2e-benchmark-module suffice?

Does this not include all the extra configuration stuff which swamps the time to actually run the kernel?

makslevental commented 1 week ago

Does this not include all the extra configuration stuff which swamps the time to actually run the kernel?

I don't know exactly what's currently happening but I know there are ways to make sure that overhead isn't counted.

newling commented 1 week ago

@makslevental there does appear to be an issue, it is not just a caching mirage. The only change I've made since this PR passed is to change m=n=128 to m=n=512, and now (I expect) it fails.

makslevental commented 1 week ago

Okay lemme take a look

makslevental commented 1 week ago

this is a codegen issue: see https://github.com/nod-ai/iree-amd-aie/pull/897