Files
composable_kernel/dispatcher/codegen/default_config.json
Vidyasagar Ananthan 9e049a32a1 Adding dispatcher architecture (#3300)
* WIP POC of dispatcher

* Dispatcher python workflow setup.

* Dispatcher cleanup and updates.

Further dispatcher cleanup and updates.

Build fixes

Improvements and python to CK example

Improvements to readme

* Fixes to python paths

* Cleaning up code

* Improving dispatcher support for different arch

Fixing typos

* Fix formatting errors

* Cleaning up examples

* Improving codegeneration

* Improving and fixing C++ examples

* Adding conv functionality (fwd,bwd,bwdw) and examples.

* Fixes based on feedback.

* Further fixes based on feedback.

* Adding stress test for autogeneration and autocorrection, and fixing preshuffle bug.

* Another round of improvements  based on feedback.

* Trimming out unnecessary code.

* Fixing the multi-D implementation.

* Using gpu verification for gemms and fixing convolutions tflops calculation.

* Fix counter usage issue and arch filtering per ops.

* Adding changelog and other fixes.

* Improve examples and resolve critical bugs.

* Reduce build time for python examples.

* Fixing minor bug.

* Fix compilation error.

* Improve installation instructions for dispatcher.

* Add docker based  installation instructions for dispatcher.

* Fixing arch-based filtering to match tile engine.

* Remove dead code and fix arch filtering.

* Minor bugfix.

* Updates after rebase.

* Trimming code.

* Fix copyright headers.

* Consolidate examples, cut down code.

* Minor fixes.

* Improving python examples.

* Update readmes.

* Remove conv functionality.

* Cleanup following conv removable.
2026-01-22 09:34:33 -08:00

28 lines
571 B
JSON

{
"tile_config": {
"tile_m": [128, 256],
"tile_n": [128, 256],
"tile_k": [32, 64],
"warp_m": [2, 4],
"warp_n": [2, 4],
"warp_k": [1],
"warp_tile_m": [16, 32],
"warp_tile_n": [16, 32],
"warp_tile_k": [16]
},
"trait_config": {
"pipeline": ["compv4"],
"epilogue": ["cshuffle"],
"scheduler": ["intrawave"],
"pad_m": [false],
"pad_n": [false],
"pad_k": [false],
"persistent": [false, true]
},
"multi_d_config": {
"elementwise_ops": ["MultiDAdd", "Relu", "Gelu"],
"num_d_tensors": [1, 2]
}
}